var/home/core/zuul-output/0000755000175000017500000000000015136672423014536 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015136702733015500 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000257553215136702551020276 0ustar corecorei{ikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIsdr.k9GffHD  ?KEڤ펯_ˎ6Ϸ7+%f?長ox[o8W5!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;FmZl8T*v (6pk**+ Le*gUWi [ӊg*XCF*Adv cXk?`QlrTvb)E,s)Wɀ;$#LcdHM)&c(ޘzkQ5 @:}ɏ "'P8[aN Ⱥw^eD6'Ύȟ >Kdg?y7|&#)3+o,2s%R>!%*XC7LnC~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ6v+(n^RϚOGO;5p Cj·1z_j( ,"z-Ee}thǣAM/+׼,Vdj:qw <ws~^B.Ǔg'AS'E`hmsJU # DuT%ZPt]WďPv`s6ƙR*- F? xtfj>PWȹl;M@v4Ifk5Þ_P]2@vN (C9yO|$UvވkZoIfzᑇy ^t }|#qKrdK\D2s&[#bE(mV9ىrſ~I=o嚲W9ȝQEkT/*BR =v*.h4(^&-Wg̫b]OBEFδW~N 97;Zp0s]UIĀg)4 B^S4t; * (DnAm_`c46A>hPr0ιӦ q:Np8>R'8::8g'h"M{qd n;ָPZGa(X(2*91nD^W0/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg ِ0A_Fݗw)Q~}% ;]/ľv%T&hoP~(*ט*y &הԟ,Tصp&NI%`t3Vi= qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIjh}iL;R:7A}Ss8ҧ ΁weor(Ё^g׬Jyachhr=yԞ~%>]CN"ہ)Za@Ι}YJz{ɛrl4ެ˧jOC d-saܺCY "D^&M){߸}t i:Y`cФIX0$Ֆִ0viqǑņbڷbb!g K#XoV甬6xڂ I &m:AtĘ5dw9}ŒEanvVZ?C}!w;YKoSrm_7dPΣ|ͣn/𚃚p9wǑ}?të{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶþE ءp.֣k{M[nMc/w&||uP_T{G7C]Ch',ެJG~\oP@gOV ]cӰJ:^q';{S^vrfZB4']a.QO:#'6RE'E3 */HAYk%C6Θ%%:ò6PT:”QVay S[l52, 5 CۈP$0Zg=+DJ%D  *NpJ֊iTv)vtT̅Rhɇ ќuގ¢6}#LpFD58LQ LvqZDOF_K 0?lm$K/$s_. WM]̍"W%`lO2-"ew@E=A\b$BrW XWz<%fpG"m%6PGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b7$66|*f\#ߍp{8sx[o%}wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKwMm[eGc̵EعuLrk-$_{$oUI y4Cz%=lRѦmDVmxюݏwx]mD@0ngd6#eMy"[ ^Q $[d8  i#i8YlsI>2(ȐP'3ޜb6xo^fmIx nf^L7>"0(HKkD4<80: M:'֥P!r"oL㣣@ 9n# " $fGgKQӦ4}Gn\^=.Y5PI dPN6 Ozځפ5) F[ڣ$2*%&h v%9HN H~Q+oiw?{.۳)-nqM?2ސv/3,9ҮT9Cef˭49i.2DxatC<8iR/ƬйR֌vN8J"iJ. T>n1qaY4ͬlyg "]Bv9-9`Te'õII kюHLa^b*/H^FFIu`2a$mc Ry;R:LڕDܓ>Y=:]t.+|PTz`Zbym gp8펠هynpEOkÈWȤMف 8o?))H;h8_ߍ5S&(w9Z,K44|MZ'0-bB$ !)6@I<#`L8턻r\K8VKĸO Hx?tythŜEǴfW I8!mR94G t̆xɯAߙRўń ]T!n Rqi3/2 0dŧUO"n2R*r<8mmN_ղwT@p +jV&P/W$Q ?I@> kV0φŜxtADx"Xh4|>XSxߵă@pE:y]/wDQ1&HǓb!n[mi3)WfsF:M"uҷos.1!뾧1%s,hQshx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nQ$tHI98/)=mͭ ڐn}uzٻWw ggޝ7ĩDRc0]rY9'z .(yYqYj5p=%Rwó^;n1z"8 Po߽]V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*ILۗ>QKҰ]Lofj8dqV+ށ*/fC Q`B 6i^SL[bJgW^zA42iG.y3bBA{kpR A ?IYY ?|-z#}~f ‰dŷ-р,m7VyIwGHέ 2^wߜޚM{BL\#a s.=\|.立=#uL#]  GE|FKi3&,ۣxݍl0HڜHLK'Cw#)krSIK<1څ 9abHl:b3L7jOq͂Ӥ=u8#E2;|z꽐fрi^lTlt␚ɓW%OVc8|*yI0U=ޝ P?&pWt|CX6o|,9:N\2x|vo|mv}r:"KiyGҪ$& hw#4_pv?\Хfm_Ov^Ҷk6j3ZN9t9ZMMM)I[RNC|䳮yI3MڼH9iEG&V Gx`u.̀ab_ XC.;o;oX]}:3Kn0R|WD\hnZ:|}Wv8|۾~Us?yWjv5>עxqRҧH-EeJ~0YIozy:noq V{q8sϷHoO͗zf ̙eX-4`T*lA" &;1;O]-wgϊ)h&i'"/ͤqr@8!̴G~_Rkk ^|9-R )҈ ^XKL =Z@ >lN%hwioiUsIA8Y&=Xu~4Le͢ }UVM)A`9Ɗې#v΀YK{dHi|}Ǘ5~UZFqЮc+cxz 6ZduL& V 5I bHb3L7h!ޒh7YJt*CiNÄFKKùMt}.l]El>Ms|//fׇ&!B ;&g\,}F)L b߀Mi6Õw7O{Gqrfz3_P !J8T<2!)^_O.\{-d)K|l1햐Z1WMʜ5$)M0Lʳsw1S^vxp VȖA+L܁,҂+sM/y)_Nvc*@k]ן;tϫyoxӻo_nfr6ИrҊ߳b*bjTc?E K\gd{(ZHCxn5w.4\h(`dc)]!Kqi4~Gvp!9V>&M!s}ٝ&%g3?E#%`N)p#5YUhN ɨڹ#Ch@(R &Z+<ݰb/st9&yo|BL,1+tkn2ӗQ7 }<ˉzRfQ*yMK*"_R[كrq IH!6=Ocnи%G۟"|ؔ^K_y׏<:n:!d#[7^.hd}ӾP'k2MؤYe/{!ca g/^wT j˚ب|ML?E3} 2ӗh$cdgK?uy̦WmÃ8ph?}z]W_M.qKbvJ> QA^"fY0_8`N O7{b_u٭}3yMaF8RŔ.MMWrwO›HzC7ݴ]MoLoƣxxy23(h>1UWwml"Ms>\΋"?|NKfֱn !qRD `MHVuV_K2k*`cKxuBG&24T}Lai 0Va(7K#ӊ!,ZDxFQO*lם>!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜˻ x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d-cI{F+P$XIfI}MvS">oIɒVT-[`HUGWWUΓ+8 hA'أ$n\e{ɑ]T֗J!%/vUH4U1y8,<]BYE\ժgUqYeXSD%dd7ϫ=^vV0*^uQVfe*3f^+_~ ϣL-6qdWgl5;+LwVS<ʃ]˨NCvYx*aqW;+pSOxpSI-OxeNE"wM8cȮ?<뗺j{N ?)d04vv葎4cg0U.q*?܇y|g;GIv'$٦$m:%lSR~ONݰB?5 7B]>gim5W2Cˑۢ1rbaG^7Uw'o UcH;EUGGvQeP]' @Qs,FŽx(~Ѥ5/-YQ<{s ]zILJz= z/K.l{r#:`؞zeMV1;4&"ICKfy\ӏi7$+p@oI>nj/E>>c+$) "oI^ЅХm`Cن04G|BS5Q8"|wOf<]ښ/y*6)wE!}R<]|4i\-ghgfo<ʚðvQ ~kamo̴\Zo&o,~]]wԷHo=&tu\͒{,-BwS-([J-;Hc$Ynم“gwʅ^aN%'x]5XrGՈEfY~V0(y/JX LߪKI(DDDFa7rjXaۃފRm[Ze\Tz8_j-/Q,^3##]<ˑfYF`~$e;Lpt#7=kM,,f9=_O&|^j} GTD[]IdidZ^ c)Ӫ;+U(G*Zb}KbskQV f8V ; ;kVϖTU&9W)7aa@f2\ $Ar]@ +q#[Gk,^MMkAႂs/.|A<|9WOl8RGOxyexbSd׃&M;>P)XV>gɗĈcoԷqlvæ9MjJpZcxp 3mݣ ?3ÒRIe  7 GdZ"qrx<\ܱ67t4sΣ +YM; `[dz8>j%0.2 1vO֚@Fo1:7r_I,`ԽX c&ښ 5pz ؿ؛.!kӧS\B/XTOH;%䅸R W䳣);JB\U6ɧ*ڮs*J #9+E>@y#{&\ZI2.`G3{G>|Zi<61.Of5rg U Dei)wbxS{RٿЖ4R-]@aNxiӇaBȚ{Jj8[b^1!5 3ԢsDmn k07e[*Yqs/g|Bzb1jz>*tjѮ1M*1͡aaЀr0 Ц|u9fk9f|/aF*ѠV6Ѭ$stiiPEU,ie .4BUa0s U {?j~\д6Qڥ=5e)7޲6(E·ò}B/͈c"׳BY @ &!ǰu (kN8Tкh=q*V*J< =xpn,?hc=iY,CJ,8M8 RI[>D*}h\gw?-DTF|m,ϵք׆li5vfaLda+q֗/>[6 pw&12y[2 ޗV.8H@b>OL~m.h#gbdhT@|ӣ$%cso9ےTl%lm%" c+B_t&&s DJ]xyXlNLc.aHc7@n6Fo;yyqo'1nҶK]4R}ޞ1SOY'9@6ڤD5>U{*\.?9\C5l%%%+G g 8vهߩJb (9irѻꉰ'$2t8FUETѴO0iZ=VZ9{c^˸X|G#~N)$V>bn Wfx LOՔtv&ʪ/O$Q/x>mz]EQ38@]y0ʪ=xiS31&A, { ٴ OƻrWXSA}xwgϓ>=pp.xp/;9.d @h Z?0'nw8H*``nh-u1]4иHh[TC#p5mU( ˱GþעN턍8UM6P^B7TqQwJ{ 5-؉6kQ &L.pk8p8J8# a{(nG\JG)XzP Xy߲-nu3аj[w{ƻ l;3VGB4"0u^9f0hXi[(~F1` ^=,طhRJ#<t# @an׼; 88Nccw(ThRh$$i vD`n[źr tHet.^@i_/ @i_ %KdenGb&)I hD'@Ю!{aq^_^ۦy?N_(պ= vnRx=]{U9[;.j7Xwp4 -i Rjx$I- -m jij]1CSv_J-λCuB`ZqRp堖vBAK-m.ӎxQbj$xXJ5m)v',w2S(e]qʅ 'm>RւFй(J5_"XRwJs=v.FbEB(ӈIm1|2Rl,J{i$glk2 ֢ N$P[)D(Mh74_3:9]BFq6BaiP[C*zwC$Ek8ɝ X\%gmj{]%XvPFrbbLr %H4#Yz s: rP׶(qm?A_'1Mq9u(UtC8e0RȥJOT"lD#XJ};f^Z%Ib)nJLr$Qq)bj] ]4B0,ޓidXcbo?*l%7cq3=n@}%2Y T%rCs*ҙUjR`+ Ҳf4$ܚ#`^'f:" 7RxӻiV4\b&rBO)^|Xvm?-:ta?6x䐼PYbT=GB였 ć,ru81ȗmvME]{G\7N'<|PӮT؟N3foN[4&Wp*è_0ƄvqD5Աp㨫QMb:N ԥu.6QP[_xjF[ bz8 RrmFlmh& k&I7՞]%A0D 7's<]jBO`'G:)ꪨwpZ4t0L>{Isf.u:x hZQf[mL~MoQHe0Kaq]QqWe/&Hr ]Lb삞ֺ6T8d-N^AeŨsL4( A$9Y ?ΨrUn|M5҇ i},T6${jѨi&?qϋ2Mia j)zjC*LE`WUCS6Y?>kaVb86a|6W4-WyH>3gdCar̓8yܦBiB&[48B\KT>k^8{/;P6/LYC Q {=HEHT"zMWu^LIkARhC K=0J62 kqde\fzp|m8I~;#(P-ZX cXd8AR1c9j>DkF- ksi.5V3Ul: .#O SÑdLȠ* 6Q"sٽB2N @%uC{UrHc$A4,%穄&^,/?ǧ`!}~fϱ*dCZ1K:~iR4$Q bBY6VDF{hayUrh0/Jր^dBDJ ߉HHl`Iv>le.k;]{Hw5H$ '?hP ;˥Ճx#u:U>7jv8~eHUÑKiC_PzIJ PraV ƼT&D~'w G[rh='M@ nGOءeo}1)3KEQj)lY'bJZ8)+,)R}dG؇`NgE\JH>͋Qkz:WJٶlJWt.dG6!W.TjZhC>T'thdϰ^ 1.PgkuUYj5lGZ 9%+cY'|%jۢDKS)&B𥏯}د [;G9nbV3aN a8;6?%t:tuչWdžUW!<\hSe(phI_![UbZbʣ& yPtDX+i#e~3ۺ|vڠTΚJBjcިǶ7u/7u|5#tsݶ (-L-qI3<k-I:_9ȭm^W^]q6̉ 4FJ*5-Q?gVo o]?=`h8CQNV^On_t Jg h>|5PڃAl]ko[Ir+|V~<,)ZKR+Nr"Qu@uxŁ?OU?3Z߮ ٿvoN_o~_j]=z|?}>U'&տ<ҟo~xԔ /NRO۟6uUg>o6=Mx)e Z]G>&^\M.^R\]M%op]IO:L?KI=5;9#b2z:JTw{i}vwqS8N'p~Ľ˛.ɠ|)zŦ}7S#f󩧃<}`)~) $U8>1~yrshnE=VNfXjhVy9oڕ&wIl۶mI3cKp2g5*2k)eHb/cAnGd|Ff/Mi:W `]ILik:#i_3Ƹ:[F`Z Mv#e,6]Hq F;T^II0c1gӛk, QG/YA/.BH:49c+X5[yV{0Np Gca>%hփKpCc"%&;I8anTf>E DU2X %Hbuĺ-BlGbLh:<dT{2dnյmr6XD_U3qEX |(YL=5m,o`}1lU39x$XzZ̫Ԛ74pOC!bV1( B0aKvp됒 #İ@6fl+%W6^hಔV:Zm 8)8!ulQ!I#fՋ*Xϫ[‹2wgQjz*$bQ߃68LUɨ¼KdFiň!np͹(Pc(5l6wtqKU\5]D⤿4uufg?> xfeyGs}< x/Ⱦ(;6󞤏Y6٨͈D ܞ>ylkK:F' q.jƈ78{$NNk`~n?f <Ny]}ӳ^)S\")0NR">IbܺeХFڐ$Bc'-*x%gRhBw ܋5AF:d!ccބE͘01c&'[I6'*l&`f*2^X!Ќ*bĂ1NpyGB㽹zm <,cnx$8J(ʸWQq'3FWv(Mi"pJQ(5&ߗDhLɹ3afhapD%]h)c]o%!y "$r|\$8eM6Ƕ1W[AmxDZaX.΀+R#Ax)ʹCDb$QfՐ> 'H//Yk׾Ɉ .ݯfK C;mjOCTZO$XI0SQ"$#naQ[1[H~QsL[jHpk/spCm2b0K&REI%mC|_҉^>|'3#ŬyㇻݚG.6  .0+ BoCʄ"9`$я6-*R>h*L>1# 5,~RLF}mE,iϼ 9'jE iu43BI_8C?^68/%R2fI[gQ$q4 MB4u&'P!5(p$JVn0.]fboiN0Jw勪y w" c0 +_/>n1,jEعeD088;weu%bpOK䋏AzvfD^vv]vqm_VR5] x#jBՒ#' Ǭm-'a?ПAʯ{wMo=~X}]?N4휍{L"Rx%ΡբTJ[ 8  |ezwڤߚcW.R[Rpmdlm#%^{}9qQm A:GSrt|6F ^|Hp4imVt]VqGۓbOn- ~𢽔qZLI(F b1X^߲to󻔖YeV}CkZ X8\#<~qB@ÚĒ(5iu$njY{_i\J/y΅ܲAQ5/e/]Z%0:w??7=_ׂ je`5-~8QR[gH ~Q-wuGg:P6U3yP'%#K_`9~V1yG2!p|GH<[tLr:H^9I$6ɞzw+W dz{B+b%OQwMa'٘8xp`К;^}Wypz(y94#DX>zȎw[x0qwYO$Wp0q3ɲyZO=͊=⺾|m3b0XPvK-8Ygt<8x̱CKg~$%x9/#Td4E]7 ug``KHb'soL:޼w:\z7co5"Zc:0bvbZicެVfKZ4WkLEkzk oϳߕI=:(y?yIWjb:9R׻IP ʶ&1 ׷hs.$tFSP+rpTr=k%ep!qDH:>|Q%OhE+`,mXUs]3C8).yo0]|*e!*EζͩvT6Sa3"ĚN/,8Y}@N|_t /tlJ+K0 yR7J*dlZerKOX|F 78;tdTȱ .+u,i;M+ 74HR5ŗ)8‘6wLW{.XY>^蟩rX§YHpLIxjV^j%ϼ-^f (Rh2FS΍Ua6"I[nIz^pl3N9W"d")8y,,=Itѥ` )fܖ,km3Iam^UOa<4@xpt#יk+f抓Ҧ\4 3>o tĠ:pY`7!!$xݬ%iY&\|+M W̬wnk֒|ya0ùC*Jp aprY 9ᱵip4Fʶ5aP'\ 7+B,*~5u{ QE1" 0tHx=x#zO9dh:3QFvk#'Gx g"W:ˤ(9x+I&yF_NkK$8Q"8+@ Mr־J|CC3w_1ˋ i3:wZ n9R<̈!ɹiI뿵"7,Yp҅#qph7uQG{Bu%<2)4Ka>)M #=^7^Vd w'͊έ:aJNaU`M/jRb Kv[Hptct+H̃7 p] #IsIr_4]ػ6,U&FR!lo،!9BV#&h:Lؒ@W] Ԯ+l0Uʠ Hн _qq^߭Q."LGĔb˱]K.Bz|~?2sdeI 9 4x4Yzep߇<,} 9LӠQ W6?ptjّW@"T"S&p*iE>4;8)*=9Oהg1;>Iqٍ"u^qR"t#W{uj5n\.f)BOšc"/W()Q.ɸZ4\S7 8 E6FeB1U98~~Nn _~qvd ϓQnВ_L(b?d:q'օ"'n"щ"—y4+˦^, ۉ}KW4P4U_t"g4\ͭgSjY1 莢83lI.E~qg^fgSX2LN~|CrM.<.# v *&Fo7KZS*`?PZ,qrAe<&=yY h@WD],'|S=W#S}p6^ã䄁x/X9Lw¥/ʣ}Gs]j'ߍ|z ldV+00OȨ:Ĉ*z&+?^#0Yyi!1ks"̤4ŤKI lfw-?lxWCn,# @,>νYQ$Wc19 Wܫ>jTAQFAmOͯ٤y/Yty6u=fn$I!=Noʅ߂Wi,4cʚҚkv2}3sC~l_4.{ϼZ+q1ſ86ѱ6uu*Z6CWn"k.jG1Xg'AK(LvX%b̿z&a_[Jhù2n$ W͊I%:1a3ފ\HԼ9L˕saxp 3=;}2(Gs;(ODkS3hʼVx p<((b775߻!"q)7NXa0%D;E~!h9WOT.$lӶcɥfڀ[#ׯYd}N!L.f (mITyY !Y2֤ٻP+:#o*_7I]׉~nݠD= @n P -=.#DxSi-H*ꏶr)g 4z$ >0Hnvvf=0 &JK`"%0铇Q>I_Ti#i}Ly qYl,o;ӃTtEWAWXF}po*s[xZutU㩱t$I 3'4NH-vRlJ pS\V5r'67GnᴫđY"V;GJQ&&D{5 BRC73o6vmjAK6̬FU"<ݻ'^+odEdgB ,I;Q/6xAE 7c+ŸТ'Rt2& +E+wFhU|OF)ZQ݃ y1z{plvZQPv@HO7^t>x D_N9r8O^mWN@;VAKW ZeN߷U}uKpep%9`%R[y= Z-!ă$@}^qyFf9R`@m͍WMT{jx W=:zbg^C$BGkLb%{cKb@{Â׶$@_v1k4A*h:C0>; 7* : 66=&ϡ*)gyf%gՎ3>!7É(ߴyT%0NPB?T vޟrL7 #9䞞y&hNn|"iyNxz~qm&" QVU骾rp灝eu_ G`Ly<(U:U& gC9(ɕD2a};%Y~ήU[9E@oV}֗_wZwd0HƊ`)Q2ZkH37Piϣ`RF V[ڒѵNѵK>U dHgWh漏D;4:*F8e#띔"%6:/q N-];$]; Zk7*nJ8MPCH,j-p%MFR4#h HڒѵNѵKUٰdxXY^J cYTsa+uh F5'];$];p͹ ܵ%/IDjIHYFD+Ӌc^*Śv2)ny,8z6INv18^AO^V_a d>y] TWߍ|zq-1wΙLAK \vo[ƽM>:^Tt#bͩ,Q5ohr."7?|3yb{znm3:\?v"$xՕwI߼6Sz8ar|U:Pu ;Auݘ"2rk5 HJ妁l u$ PSH` >F"v #K P~I`_%(*mB4!lIH5-v\j'(DX񝫉epBö߯@BkH3mTM3X3}4y;r~ՙVJ̶:bHcWggԊ7h k}?GsTTh|0¼q=TrA7Kmg ׬f_f2!d_C.Ff - 1oFRsY h39n|_Kr`+(nݧ@~FC†H *@3-5-CJj_d~?N. b>Yu9W+bx¢Ұ  TX@El*gac 뼙._]P$6d."X/irjmkNX-Q:3hJ#YŠ9M7QDQ)vuظNP#G( %N:g#Ҹ P)߈m^lXh=GQ̤ #mv pnӂ" EAls:Wk3]DGbz`L+eȚ5vtw+kt;4 i ^rnҹH\{&3DZ9 [Z ;-z1gڤZ}ݲZ$)\كd>2Cj RWBK0=ZR1YF{sc1Ͻ[jM!RH"SVLpZ@`&壍 kU 2W"gy: y$ (e+ )WOHL1X擀hX1_XʄaE=6x-DuDk;2ՓُFDNhEpa#;iL,n});Mĸ|Ej-ltH0^*x٨-0q9`="4DERMY%Oob\;0 pp0 d$H6)EC Nwyڧ$ARRlNJIIԪR{25B(6΁a4"NJyi"3**gS5:y͈7|\: B7╴0ũP1 IBg-!V)h`#|}!!1H11f[ pj5I=犃Vpjm :Bq 6Zi^G(@['c@ȸgZwF`Ca N3 6 ^ֲ!%'A}HJdEw=81oUuUuUwu!aie%k@b$5Q!AJ 05ȅ1sPYA];nR)Ol{V=r&q4StrIDT 'v$T SXUE`i$ ABY9mF=Y9mͱ6 bRƥm\ #ƕ}mgu2'L -Fh.Z@H3r2BAyg3@~4S5KT0Tы4{͹QOqͅfOOHjXK&ZT5{p -Ӛȧ|&,N1 @.bB8U#NȄ^8  ֙0rpDG L“ bNÍs)~_#=kͱmdh LcY&] 8ZCҰdfT$PE)S`9s04Q4:ZeYbhXAO_ y0< ڔzectnY dD [i^Z d><<ga\ӞI#84@Fh{% ڞI ⯤})QjmO'_0 t#LD﯋ \ ׳e-0TgFuU)`z$@7;d?[KQ$h$D)!pfqFG k9 OHt94%".JtIl*az*a$aHXsK$Zr`#čH"&p*R?IڸāIqltdThb,$pXPa%6 qO-Y{*DCgc{*13x1h'3_WJ^hXge쏖YQ̞!AI{zǬaH0;4h EO86oDfO`x37Zҿd6 E4Xbf!`zYq˿,'V= Q++{-{l emS7lay~emWu?Z˕6l$Ԁj>D:߮Z)17JU'qcN33VKuǫĞϯjQO؛g1i>:xa>LvD#*ӻ Lmy6`u( xo`C:ld&=4t$ET|RIwH)Pn'6X{rc&%#.h͘en%e<97*) \#J@&j nO,{kLUB?׈"#;&{5lXwA\{t>΃0vtY&0GJH7Jp6}QB ZKEVJx־lgo<,Pñ.<`8=ae:4*`dy|v=.^/F-PU(R6`LnǟX !>;2`9z1SR?V񕃖!2oנe&~Ʒ/Qzܳ`M9WI |q~a `e.% D )rږ7ѠɬDCrzp7B|^O=(qJW4vgtgi?iшw65;6]/?5Ih-vdOtZay"tެ4ssk O*ZIhy6g@>ts%l^so Çz_q ltTo&]]x[\peQlgޯհj ,DZz[F4^ϸֈkX N7/GuKYD`>HEWuB"7sg<`կ "II^2 yF5G/I/nVz}9Oj䆲'm9#s>G8IM99tg% ٴZ5N/g˶M߹NY"M/W #~|^]Ay6.^}69C`~! &?AZ5tqFL5 wValV) 3?OfmɨJxUHBuaD<[2>ǿU[ 7q T񮟆gY]# 6A:c}ۧWU-?eN#m&᠎_LqA-m~slsR,s&'@QVϼf 5"7dd~?O]о^?\F6O}шS+ڴ|?7ZXma\2 |z><|\} ^W0-`RTi=q&[Cc +ZY0U(RQ<9GJ꿷1A"WJ6=-/6a*aEF۲[L(iodWNy2XY,D!^agIRU9\Pe@p{`v{iRLcͽu} ^]D *z`Ugզ3fR:jzȚgmZH*Z((z!6(]&WcV,V| $+*A.J  `Rh#dqPWfcvBLyz_N0>\)^d[EDSUn5.t"6^]I$Ku388l始)Ē _J%",$ yfزt ֫MַJK y4%&Syq;|iBD>%gȌ,ƓB ˛˃ib&ʘL;޸r0l^MW/;Zz$V٥bmWk~X܃d6/R Yħs/ Ć0Z?>+'sޔpiELrWeٌ;o c&,-z-Yy6+^y]j^ I0x8-H*7̫hS˅(bIU  k$UџzAzvYvQ sľR1,@2Gv4Úi"c;|:i{t }K%FwT$CZIHP;FVap ,{l20BlberjtqbR}IIB[A%xu<ZmX|N%_iZȣ3u͋npm ZRhouS2 7Kә-C ˛G\kǃHl~Ay,I|TOo^Ϙx\ <׎y "Ǵ gҞ&ZGԎH CMڤ[JV^NSI6z=?\\ǘڬ_.tקOߣ|!O Yc|!Cg] nsߟ˫/KfLHY%i<)X<5vߦq~H 8/[.x#w\pc}uH_M|! yq|-?ȃdB^o.fc.UeiWT0>neeeXc;MʗN(ٿ3IR:;e WU?-ga]7b@|؆ _Lapc(c<ЄN BkcELHDQ!ձcAh%=ׯp7[)5xtWXAy}Sݤ 1~dkw~p8J瓸{CVYQM<_!qpitlecz9MNݏ1l a#VoQWRY!/nzKLJ{.%X$HZCq?\l|H 5RLR+áđq$zuwU_Qn?> -=\PHB5.B0 )L v&R?z)¡CN\Z@QNq2Z 7kբ%݉:@H#!}Eeav ?ERAӉO+3B>CQ2Z[!쑐а,K2ebJV"Y(J8gҧPuB3_ ^2I;ϜfMOU91;R5ڑGbbZQ7FGg9B$NEK&USX@D*8|0FU\DY9> Iurn3rKF#5I{qA:tR"3P+ Ҷ8Z.=CȆ-* \늶3AV v/[7-nogӺ4O9WOWOk]l}0Ȼ)"!~WO?fTl_qkN bhԤd1$DɈ^IB"{z=c0ϯF՗7h_z9v2Vל |0D)/Z`%h9S,kG6H7Lx 2Et8&j/頱a9G8,B6\?9rHU$`+ db.Y7tذrxO+)N7/p>vMZȪrD*jզ\PeAvny]xOKQMDU+Vf¢;y]xOr2.YUY^$O!b#$UI* &mbl1ޓoO6zM K^^~a:ZǺA~'!wc-~qi+ȼ5Ѻ +vTcqdr`5Uē \V'+Ƹ7xWwES"NɕX٩8;@k}`oܥ/;ԟS_[GZg=z`իqCs#i&2Uʪ I{" RTi ayb ]RlcÜ "SW",*cm(UtApP\+Z}o`! 0<چ'ۼEyv]ln<Ğ O UY}0͓Mjx{ֈӭ;Y_ZWۯgA$hˡX9$ZLJe*EdwTޓNr{k#>@-Rtи ApZȫwE~JApD T)8+yeC<׮Ce,[at0RH %U8I/*''"QODl|pqXY:*=%ua;a]m}<|G.?Aわ*`,[i}gJ@pK)Dr2Z+f"Ɔ-%q;!**\${y0B&Hu"4(,WvHg,?ȱNuKvܜn ~94ۻ:0j30P":yn$usZG/ f31`\;`v{2WP31 jCW)wjCeM78& /˹(bHW8~}La8'}{ZJA㢝Z0%BW8P[%oKL2!eMvI]rZeNnhUC^cPXU,KeZ|AD%e .Tꄛ4G$aDCb2 9ASD(X$W*<~eA5 FJ=PI˫쌓B0k'pWt)W5N9?DN&HDF<\9EOl΋bx^_|aQ]piep-YU2DJ +  AGXO[i4.G;Edñ}p$U%rp]ǟﮚ YVD.+b;eU`Zq:P\ZqZp1ϮEK#jp1d^~[\1:5as^@d]rpqy6=$:SZ[!n %-ռKkuhWCorDm:zgp~*;`x|Dk^"mh=Hsc:ae)~Fky WLj*a/oF n65Ao[mz(,fӖMsiri<}ۈwQ*a櫕>G-kOxgE5R=_>U Bqa9`pop;^u98Մ2HS|17_ q}͊-8R.@濟g.%GXV}|3x/fӥ5|5~29.®q}t?Ya{_h7s?]~y3ݢZ7\̾?Yǃt >f\LlF;0}b=nFy<~hhb6ò?Ζ ~"ߏY/}3/߭$߶_{ƚIg|Wx7A-e,Rue/|=z{qS|wtHCx} FXOI=z~n ]'P$Tw'%*ijcf_E?wG9ޗYڠ.!5'pǫf<{c,7o`wqsEqFkKp. |w]ЛS}?J2UQ>,ѻ!eۇ.`;slOxssƓZ A8~ ALpo B FPwkd ҍ-R֜TWDVQ##DW});e[A)ΙsA#7bn.o%G"Bv^m&f>d n&/U9&$͒d "_jφ,ڿݿf^/~PrF> bhUI|P 4sc xl<DT,_Fn_~Gs49ٴ !!nBhK0긐 ;&i|1|3%(Zn޷$m5J3m46pz

p*>im3ƃn3#y^ᆶ)ZrLrv5qC i81Re H6 $BN3ʡpcj̿]Lb! k&:!܇-ʎ`%OPOyFM1x5bDrDrZ*tZ)Qr@I3gA =oVzRjjGKV*m&Yxg *U8ks=$3 uaQg N.i~0/6r?_tB8"c<8|3V-h>5|PaO=URO_W7HR AJecptur:G˧-TZMa>ykRҚ ()@a+(yl>-H};Bk<$nx]R#`2k+oASn yzf253hM7i7 %,5^[" X]k%LFȣ#ȗjהOrr"2*2}ɩUf}9b^!.hNwHo e~<qBȗ:cx'0%2)' PqVi̋{J)i 6532D؇strEt!_r.E9cMAs#2N*F4R%ogE+QLDsnBKԁ3~=?!ü FĈ4 < B #ُ R.Q<(>RpbhH>8*kk+ZPcCY~Z"X鸖S12 DsA!آ JY{ڂ[Aaiő2e=ȊWa31^r"X-a@Ҕ3V )! WAaG g*.牭 k]ЮZU{TzXGUޤcsFEǥGVD KTL 65]Sbf%/Cwjad,r5&,p<9qgA_z/om k<&XmФ vÚv%g`Y)j?G ϗ$!+g,r护ح.d5ay魙FZ*(5NHYk@ D,C,(.i/&Ϟ=vg/S<4#9wrc_qqO3SofׇM,1RO $BO֧{( ?! s@wy <" 8ѕ&ʕ>0BB4:,Z:/ƪԂ%N[ < 6'zAd " 3X 객nરVAto6kx: 5Qq8a\NڔU0qo^ԝ~ԛOJ.NgL|qNB|q"YG8+6s+ /Zft[$ֵC$g!׵KA ެɏ.H:|"NC9IqʤIGk2xEԧ s xĊ*M,х \n`|18灺8|k<&e GmW?V"oXQ12'\(*)N%zql΢?[/焏2XrD0z#Q鱎ƺ`<4me{WxȞ#y@is 4&U6 V\zU4J[d?d(@\cT8GE3E:>_!ZTUCP\x>1Hw !BGx1aaXŹL(W[MG77Vg/ TO@9M5-7g& /#P1t(|QX[T/Go(yFiZѥҧ:.K͞T?[o?PL79`dz'#}&D oֶAn>@7BL 8c۪ϯ65j_)VLF_hAX8KԠi^/DgXK4;> F6̥~ ӏgkݵVCH䈕H;Bypq&m#_M "yv"2- *Ht6=coD zGebGC|GAKkDY3zkTuzݼo9tx[*IdO)uClnŎ8d&Tuq2-#_e9 ?bG6?ʚ:樔kx\siX$h~kW{y_UJ/yMv5Mcw QF;Uÿ'TJ"2Tnʅ۷K1Mc DL3\LTK9IFdTC%Xr*a~'D935H酗K[An?d 58/TyXAXߺ |{o=r>q*1t: 3/0vqw;|*fGo #^ꈌzGKGxE醤B#2*΅UhR@ӓEhldķ^(ڌ(7hp_<}ef ! Fʆ/Uv[[z}xmh#8K "fH6sJa3%"a'btr/1%S8dLY=&a.Q/ \Py&؝2kE`4ye:p#("c;c"_r/wͱ26X;!=5o COچiFPۊOSOӺS=zGNh _nEWq/կ$Y;nm+Ck?c]OUU{:ֶ!B G{ ڭs} yh "_TQ-W>DSơ%H<aTXG>Yȣ6R*^vAi֘6Dg˰_ YR|&G|+TO~ 5Ձ\=]bΪMAWX1D}G=Lvզ4*:ҷ>޺}!- =s=Ax`,/y.q "_bI5@#*DjͿv=ҵݬޣ |1,=)|W}]5|HR e:xN}1 4a3K)̈́v g BTfA] \a7 nK8;a ;"2pJ3bv PеV2R$kLk7S,(5LXjzS|ͼf8'сz6v_e_qŚ3R?mJՊr^oSJy 0}aRkS(̋g-_÷ֿ 4L;/AZWK +K,8*َȨBgj?VheF A_ |úb$e2ZZ0tDpkQ%A|Nl\] e5Ѽ> @r\DF):]pÎȨ9a%͙=0zLf3 U\SprA`9Sʭ0VDO*o)+<];ENr_T0 ~o_΍wB;ިN&xOq rPkmZ_w.[{eɕ]u6Ͼlw{[(@Wiyc$}rQ_'AcƧZm=Rv{E|&GK [iQօQG+^"Bera:=M) 3QlچrPDoyA%l CA}R>/&޿ҽ4U1w3OҸ/ !$u`y#dzNow1n쾀OW_](^mm5aEdɶRnG`;quSă5ZӠZC1mv߶:\rΛެ ̙f/"|\5;xR ׏\[ipИ`w,Al΄O]Pphja?n E>]bߕ| .}Z݅םV3˲AE,(ؽ|y.!2q^R6OD8eN6_l&%$sES4AܑʫwIX3|,#N FQ F$i)k՝ۿs_:3AғejPõMRrU93M~AAsa!򆹰 \lL정/$^֡`7:Ԝ]үGBEddV]GrJDY73+D.W<{oD?Y[v@ _yr@x\#eYuSS_úo,&!np5ȼOm^?|*ڷW{q$pwew{~fftcAOm)Iw/_QeŶR"QMc**XUiMXZHѦ)UƱ`pjàjz%2ԁcdhpϷzi8%vpua"܁cd08Y PC!@ē/yѧ "Dz5h0-@1[v^/Az9fd*Y:Yp,*-lpr:wQ҇4`2Km,A'sE*zAU 12qۥIaWq<8F&J\G<*ʨǸĉ}9 xΓ\r˸\^;EFB: ~&/7U4ʭ,?X26%M*EKTng3@(J7-_մ$i5@+ *-\ *5\Ò<[-nؑAkS4k[NK,Oqgd:}1I3>12q{nL=i?)83@*Ҝt8Mbc!X3itxّ-~{2^;pL6|oO>:|W?O>yĄxGz6JS*Jrf*8B`P8o8 12qFh.R#ƦǸđzaF_`{e_ ȤCEN8%ZMKa&6WGa9'T2X9q.t8bSrIĉIM(}@8kqbt,;˲12q(yoH{9']\_jY*EoϑVPo١ж1+f1+ 1 =p3ֈP {SK<-[1sh؊b tJg^t TSW:O~qR^Qxf3aC1Oj/1mT pSObFre$|; /Ì0ŸC3L<20e3y(:ر 1U.VQX=Z50 >Gm$?vo[b5,Dw4α#qoiP,dgx{I͋$jx;xEd1o< P]|;Bxe(=Ba ՛<>V#5wX [n-pj!-62T<۩2*w*N <|RyY4r->qO0s`n_ƌp!ߧ nÝPhǥKngzE (`Yg5&nʳU>n+~ܗU=ǡ- xAk_^JNOA~1“ƎHT]4lgw얍\ Q]l<5\ `Lhmς0ZS,q(Hd97I nX=/ř;S[Vb5DwS\pcA iyLiNd T2:.Þ7ESCD,^bM=),.[£sUy8zQ (`,GiLQm:PFQ[M9ZC7NOc {L%Yd~,)Tێ( Hh*n$ۑN&EʢdACZ?=]۾eNdN2"ϋjڹ;=N*"KE pɰA=nSCWg>[Y5wt]?b_i/`3+R>/u$ê׷CH6~˯ƌݬ^𻼼Bpƹ+o mOOc LKFu8$NSh (_Q04&nxQX-mV3SKP3r SZש+-`Dܕeq7K~^mcuU 0z TwWy3:h?^ǪLyZZ:XEJӘ.f3V pGޏoEV@:mf8 K*W鏙KtղtspQc9Ÿ0[̿eiH5`=`0|bS'_h[ՓaeE"~H>OnQdTzˬN͞qClE jyd] ~$ogƩԬO`Y{TW/{|f+2fO׫dMJ..e8.\X\D*B&&QYcL8?aDyΚ&?6ixgM >Je_*n>_ 2(Iv|Jo2lXKt|-g%y=Ydoj^~\I:K_u"kcy f[sqKZ ^ 7_8ϊz6%џݼܾLA|=Tqmţm/ʀ םsIl೶9nUOH \,7;%b~q^T'SGrkq`0s<գ+[3O﷏W;kw{x\T> t9[%0M̶@ 8ݦk*@a'y *{^~LH~7a\7(n.(=+%K0Ě_@S6Y70eP]A/Ҵg0#63 LTjfժ,Jy݂_'wE E 0];9 k;m$ݡ{໮{gWv^x6hJ \FsbT:\GӅ+.N1>Ny|Wa{JIKI(8ǫC9[ʝ; t FdF<*4m = B2&S 3 ,"@4 Purq3*J pHp$*dׇ3Vgpt'0miv£{rw>ʧF;U \TzлSSIX'4{Cu6+f~"ۿ+Ak\fb]upu}u~*v `lm2!N07tGzWbԕ%`_G$DNDB-elIcۙ!_կ fYo6+}&ta3}$H(r; ~l?ǚ!a3ICy3KkXR1[_46ƻ= ed,| 6imzn۩o.#-a7(- =qG:Ƒmҵ ][?[FݾQ )oT\]kkciOcr)xPR$́5N2+Srnk,>@6>w+^ҋ }4 ̝l6e10_c!(g3 R9x&A TXKAՊ >nSpvKl@w |^,6M5jt]ܓmKjYr3eMto fRx9% Ҝ0jӄ i+XW@ [WFiMᬫ|?\Z s_-ML cedɑg+ˣHlJ:v G<$_>/.mŹvuFj1:{ۏ;ӨuqXfz nc/uwsl!Z6 NZ<e٤U)hXvp|{ wI(bB @$a)KAݭpUVn9ތs IF!G<k(6Y7+wURQLb2%em͝MӃLҏm r)1<;48y'Jtzl\D+88OX@.jXʒ ET hc@e_*>g?p@R㍲7[=mxfWQ>QED{x]$z+9gOOΔh#pnPENs2;rF]7Z?L鑪MQbf1Xm7H(ջ?fg9JS#vg@oFh4'LCln$j`nsP_l 4 ʧktHQ<1C*&Hȶ]\S1A)FQ),NErvhXvT;@>6sF6}u/E9u%Èo^sڋ3_.FR;_"oQ[ Fa(j*F3$: %Ec:u !<;FFnwK^6}b0#QqMI[f &˾nSٱ dA\)q"d㾗u4+w%͊ oq0O(61o剨{}?}<>ئg3IRi e: "$c@:Gh'zƆETAG9IFFIAs ,FYne*UP1NViPˀ&kX:Vg<#K6;?H mc´{IHل) M># GD[ѤT_6n[*hc.^!Ձ qQLL ÁVWrX[4&+icMK ?"(-8d|HHKIyN6qK ?ㄘ_=#Y Z!?wn߸Hhz܈E& JFTM}L]۽-7v:QAE6Pz3Q1{yZO=5qG&ڊ#g 5^zͨ4{ʠn9{"'Zfz}8/可/K+o.˹8)(;2u:ɯH?Ǹ jR&+کQh\$I4sepF0t`)K(4B<~r}=sO/3sj "/bW8zg#;z=.GQgq'qt0 f,⭺{1LyhrjLXlc9MٗY7R>?&:r40@oc,ʖG'<&=h۰ kq$,Qltu[[E2VWKkQrsqio>p,'V(X p V?S|38`@`msoc8[}a<4nNnlvk5ۚ/)gN+e͆ 61db, ϋ^Gv_UR ގKAPN#O Tu[oc,2/bO^?l.Yc,Ҭw¨1w=JwcXc;ʧ*?i (әEAݵWk f(V_^gteYnf//Aa{?(rJ@Nn7rE0Ѡ7͢uz1_@My˜xzw2J)`|]ٛ;0oW4f3ie5%8k3!^3:/>(`axsC ޺SAήtO)+)~E );˦DMzw$!ܷbZbT(.R_ëzFz,a|lA7-er4{]EÝ-|]J*̇-曗\ y*e3OM "*ںxco\d/U1-tW+`E^gm/11!eLXό'!\Ϊf'2$9bO1Zks 0K/,TО8];{f~ Or>52EmzddS5ؾd/9ƔT KeʝkY˲pIÑ)ħV l;~?S\A(APMZR 2GIgR36tDciO ]c[GW:mg.{qFG=v@h;bZת^-($5p gLi{ُ`d{66@km{jGTC pq/ Egם`Oxv*-7t<=Rq[|w\>^fuo`dʣ5Hy) he ^.]9Bd ZYo6oM׆TpTy0eԎNS;?;(BssGu.%A?_s`mF{B3+L:URAј;N$85 ՒFs^rF9/ϲ_M--ї~?"O'L(q73+zXv"gD܂1ӎPf3/V8`[osW?ݺs!1Oq-V`Xx@nY$2coIcgU"T.@Νs48۱.3eZoEpL zoPFJg8 f ^ Bg Ja0k#)TD&Ffi%S\h I$0{m 0N#hԛڄ%EpNc%\#mLmy D[A[kg6}5cݽv +H!'MhDq=Rz, 8G 6I@%rʭ6k,pdpdzڹ$@gҗ#.{;OxN5ËlF1/W,10 wb_o6 uO$,R-7]h>_ SjX -J]R/23ٯ,;[\NjQlһyFVHJ"mbKL'i]fUzd8 NJl ^{5E$yur&'EXdߓqbqyh*O{އ^dV').a1z2uޚt%__ZA>~WvRɠd 0O-V`(pD0HzSԻRL]Ҏ> .{awO%(gkrTap9`"ndAaX:߭A3"u)b_VYŋzd[ix")UOQ#Q婮QFyHP`Y}P JE*L ijڊ\ۿ}G9ɯ؋Nh/Gn$w쒒{](v"!kFߙbM1BTwueGy thJthJҡ4 M&zX<r 8M(rnS Ȣ`;X+\$JK@!;v*JdGv|f~83F{Ϸ_Z3]Qپw8l+j.2P_^Nƫ@9N_?eKi-֩4\.Dy qRUlPPJ`}~;vYQkta4$3L w|I* Ӓi1$Nj}MϳS~9V[Cq Ik~joóZFѶiZno,(-vaDz]-Ἠu`YXg.׮X;4QEM;D:ch$ lF' kA" ; ZeQp'V(Բz N!PݾiOHZR <)Έm7avMwZ} m,F 99~ze@vaݿ(gMbL}lCA<+(u}0)k+M?d1T_'N/9+/t6&}l3W+(qu-?+)*V)3dʓl`|5UmB.=Y&xW m.N8]X05BY#V=ZVX9ώe~϶Eyg}d݆[a&Mdd1 hLP[,AsN tbᬠT-A0sѦLށXz&u%]7şTeR-8g"T|%;iEq$kkI!~3$gXüq0m0M WkXAx*mU:᪑*PPy*d`JqTɗycpKV\hN{t~s*k#2Sid_RAq$Zˎ,uC(`^ 2ԇ2o`ːCP~ 57̧ ܕcn[Fy6=[Z BrJWBorJ<L*Eza 4~KRfv~:\9 0N&iv-[_T($*RKn7M1phݖ\X 57Xñ&⎪&Nui\1Oڛ{ikx>%XYÝbQM C(T`ނLmHIUgʶ4u1 2 Sri3'0B ) MH0/e,C2|x%d}PyjG?,}hW`]c$y`C(`t"mʄ'[9Q8\!0oo(dA7!o,ʐAX'z@W{9a2[U`T#C(Ta~jK%9BfX#4Ku9phJg@B{ sspM(1/PyŪfS=R04:f!zlPẏӔX( UB3$6&9pZyǼnh^v9P¼wN luqP|IfIH9;SB}g""&(E PyS])) e& \c62rPb"' DȻq!5  PWhPLAc4{ht6dR#?!&Gx)4èB$lc,ѓeTj$Cj74o DL3 p\AG\bq]p4fm("alRcZiHh|ywƔk'@r*\1JHa0$!cr-|A <(IJ|fnI*RF1#S +h `aY+' ˜!zL_Z]z|JF(ӗ dr=%cBn٠gYx53>$Lpeǻl?GhxvQ Igi=[O3g2Wk1>+f3dƒ7 a@ '{5 ]jY:-cmg37U.S%Qǯ4%ѲfNȪzɪ/oo=r%c ?$o6{"+9, w|ss [Cce*ʊE3]pϳOcRT+a:+::P[{o*=]wP} !笻ag纊lLK&xSғ _#-PZfƃ@n"F᎑*v+mtb\ݺcpiBNP D(gRt)2YwgޞRrarUS580{jm))Gn[?SQnv]ovŻ}2JOFL"|Mz w6Qtu/_Igߦ+@gdiC $)g Qxi4ZЁep ԚzȅƬR,2Dm)UuZwuA8>$?X#w@u4s`-xk"&DFE+=11]X彔8w+~C)un^nl.1:;JYCi4qUZ.7Ų ˺S,]Sq>}s9x=@uW6{~8*ow7£]<'DQav lW-~:2Wy9oeH|Y=uQ`Q:~oJ0o.'-VGp!&I(1orD7D@^(4B5VwՆ&? )\>*%H_ pn?4@W,jB9;|m6-<#J2΀=*|E()X{B:P}(;dԽ&mn"\%V^<* ;ir6׀GbYEMoa $-u+hI Lh|ʲe.?н%u/a˶tnn{wOࣶ>{ޫhsT5q=4z$4Nnz'r~7Lxr'n-׊;e|V cYx.B̜1(pH#%]hFi(g̗<^[ݓ6s*1%&ՆvԔec p*D&$eQ-p=t 8>K/yZF?_9 ݻ6f ڬYeFi[/FiJƓ~.EaYSWI\Df/Q8f RL+f {Y^F KP1@ xs`/G/C%.?)O:uιs3j`#vkprMD0/5 Z r-(Ñә=_b;;J,[ov`=m/r{U^@I+)Ʒq8]irHXOSF"E/\TOM=9U/'"Z ro6YG Ůi"`W!"~eWB-> .r/Dl\f4?DMx )sOntQx8(!۽=PWm r5$RS߽2X s}C6\,UwŢTm\Zi1Sfr=HX䲻ZͤSb Jn?k$wb@֌-m”^G8Z>=Z<>z&B,),g1Q9>ϓWzd7S h?^]+6I:RՒ- fXw3R̦<t>mmFb1@ϯ/ocV]vj^iHi`o64zxLB;7jB^m]XV%яW>=_+Ϧg蕛~%!WYs, lkhɪhY\0x%ǯ?z~w秔g|vp y QW; }4[5͍ҴryM;f]ۋLI>P7NڒΥgL$i4&j3XC-JuL^8}^)fӺ6+o= )^{ T>RJm$Bk-x!IIMHIe ;xCb< r6`..j~z5`ww3!>>LlC3{ϰ c6 ؿkK<%mQ\.&5ب% 7%Nbu L '/}*|W>Q;ԟš6\>KsLf;Gij(Yñ6Kg{8Wiؘ4U!@X8rfc vxP1E2$<ᆵybrb7~޾Hg`tmګ+6ؘ^ggV6XH>]bh fK)9 Q %6Mgdhe޶@zmʴ + 0u_k׍o-C={pfmy)MWSEĎ֓-[nYU!+(!fc×o'ѱKlt< *=^j+y V9&r6ȹMŅudXDcZ)"VdUB{l\t9a @hjy ݷh充zqz4"32ˁ΁΁~҈G+&t|g99999}p"!Ɂltw|w|w|0,"kQZT6E/QK-QnKD-QnKDO-Qǐ1~ Cǐ1~ C1<_ׂī-rcQCyi`pFO6/ ]9gexPQC{NjY?2Oß`Ř`Fn`Fn`.yS/\6G M Gxip%֕;Uf)Jژ 5"1kE1Jǜb`XבP5 H (C"ШA`vI*әm0RS-yA)ıH1`&x[dMg;&Y?A.hځ(Nl6R UOR=r.uXEL0C6$'јI&rl5ou5_崠c>rŢFLaAT d#`3L@*B>"B gB8 k?v5wҭweozoJXɳΓ6Rix`,(U2xi23ŃzQb* ؁y*rIϴjnjz#`*aU`Uty:ہϵsO zl(7Vjk",Q׍Io|=[ت3xY^8cCaR~ePڒ-X*&zG 'EXgĦ߱<&P iÎ0CYPT{RqXt,>x LqˀY,C2Ncʠ2zR:a=zΒ2$Up;o*j|sWL'ZI}Z``VOY']a\hɥׇ?;|4۶ v1P/%,.L8?#\ګݕ~ 45: OzŔ+ڨ1/0fޖF\LRAA!+z"2?.j[faqtGvc02rabJ*V& wK^jP#:H*0VS{x_{AZ:l{[dۖWPGrH||z -ПPUob}i&UZqV5VWpyώ@}}QVbeEAPA0-)w\a>jr G5.|]$RϦSfMF0. :Nv{KB|+.ş)~ZUHKE? PۓeA/:EUVุ"u;3zn~gSI8~_ŵw+;rߝ{]^qGFU!bۢ n{6jR]}IR>3'眴˳Z$I(| 'aP*f]cms봷3l=^q}$$^. Z%߂~U9%|[6{ֺ8g8&cD{ T7X-i/9vR2|Y>pJÃ&ao[2-_X;eh,캪O޹($"Z=MẄ3%h~HKmNaW!1™/r!p۽{EM d;WG4foeUP`0) nX`ūluK꣥.Nn'pްpه04N_>uxk1ܵ~[8w\-m׾(6p\7vtӗqF7 8Se]f5ƎhyS xlpߧDTN9tچb{vyDKbNp jz䀙]{BjhśaḰԇQŇM<.N&yIIX1h)c1L[}0wpU1W%ZO_ƌ.t5N|(Ttx>L^<+Oꄨi,C)LOAq2SLU!>XJ䛣_F-d ؖ U`hi)1HWL|~~kQq+MAIRsEhiAQ)H)c s]DՃXH>YG5%)qJ7Dd%0Xiwk~dh𐚝qe}2(\(G jw{QP(woӑ [ix_ *3 V,\{ql<!ս@*ҁ$cז&Inh|H1TQ;dT;)iii5S2Srǔ1%wLSrǔ1 B=1WҘY"jtpq$,(Q!3kƚ]#K:6s6]ӣcvٮazx֌Fu QFv&`*j Tj9k \AFZı$p8Enڶ5FDVKM)q,UvpuQRk^jF5nm:ۣWČuϘQIוZaT@r黣s 4<>]{G j! PGRD2$rx2#w3 S e3Tƶ@|.R-43 H^G*#R"%j"44R%pD0AP8@2ɃM"( ܏? YyG^/3f݅%EpNafsEqϹ@,Z[#2,D^?c~WiFb>z)FaMU 2Q*CSJE19u:3RvwxŷK8_۽/0sߖ}.`u7|l!|{.fjZFdIMM&{/w'tmuȦQת^P8ŒCσQUrbR?\ J?]77c)P_u.<S|'WB,55}_W أ|^;MNYե*RK˅?~>}绗g{/g/~w:;//AeR-hB]j)u p ߧovZfCSvZ0m.a?Y;h(@M~"WCxѲFL)#*1%9T=4 YqFO6/ ]9gexPQC{NjY?2OO0b: a(hK#`x4&sOmt٨kL{H5^ Q:5͉H?:X&n+IP ߙ>,-PG,]cCL(fRJn6[BQQY  T+Aq\ $FuSHk4q1vt7Gko$-_&By_ɷ[Im KSKjSv7&b,:a9~L&j*Xj~X̀*[#ɇh3a2 (d26f&$RdxbGUultl&;5ua[QQ8P6P/ke-嗵_ˊkeͰ嗵-嗵_Z~Y/ke-嗵_Z~Y/keuje-嗵_˚ougH_Z~Y/ke-Ee_VM嗵4_Z~Y/Z~5 ?A3ytހ4'h4RQAk !E=|6|HC[Î l4'5 &*y p%"OEdLr7cuH;'9LĸYdom A:ZܐO<undu |"a$<Ѝ:ȇyMP8Q/e"Q?ZdW.B6 a֧4tWQxQw>C/ IM$;(56_!z fۃmkD7z 5l*8uM0M}~~?X^֧}g6zqo%BOV^]_(Bcߝ Iαd͜S-{Ih)oz̍`ֺPT H_K| m&to4!< mMv&w;${+ܷ2s)7|Uف/tп.>gwiػ xH$6q[~&ݺW?y7 >d"eݯvAnQnl/eLv3D-:8%sGiGCZA!|Ԝ.O1hN{}ꂜM*7@i& >(MŠ)XOpu3\"/ ,@~ct`F W)xFtT-bi-cZ` `5i 4zt\o N"`Po7POpk$9q] G/eQFi-̐ria02}'{W{T悵eM=K6 s!^RС(p1虵!6ux4O!jZ5&) e aDRsAy)u~Wݹ{}k5Sf!'qt'CRyʢx4gRVZY0rlhj8'^K5ݧ \ljnwL9W=;:N 1L42րLɉykC蚡bЩ94B Bɜ9qT mJXˤA*S2O9+;ֳκݎ6ARQ=0 VI [ 8iٱRx)C'&HiTlPRU.U2=j7ӻ(%,F~NJZ,%bp\$MJ`gjuŠIO%j0qcU0"PY .rYK_|5ʘoC2.ӡN ^1u&τ9n=XD,Zp#trpht#k&; rgu4Ћu&"U>ӱ 79%G,T;29 H4!)7\rE[գ;&LBرv5_Y"ioi<6HtYBBM6Qma\Sb ݜj.F.l4MI4Jشb2AM+~i;?k})[-gE8xbYk v& R-6F;R)ddlќ CJbRjUdTspBNhL*Gۤxg/SJ(3DK&ؐ4XƍRIE {HFpTA At,;w&>gAZ5.yLlPg-ȘR"?Y7o>ҜScO m9*<8TJ15''LM.%]6$$1_i!ۉN3!ϮMXo5_n .y:+movh2CyEzƍмQCxׁW 8nlc΋eիyBt#])%R ҺYufuVY+A=Y!cks렛19(w Òz 3s88W/Y49["ړŮFrk^,+edg-G|.?^w?3+R:ڛ3xbދ_~M˩@eR{=e+ eraҷgyWKԒlè7]ok>&SnVWS+oU(V_, ׽q/}(vd{|ͷ=㍜o^Jk=itEZ^Hoj'Y@[- K %C`L,%[L$JP}sr* T\S7|PtgxЍ2-!!Yjށuq_IWETyCp+xQsr6iAp3ыG.$H빜ד?>hkYNݸ^fB'2$d.9cbr AjtLRDmHdȤ?ʠ^? fɷ2: c@ ]g`єrO!i EwoEG$ [VMH&T>soH,ˬgA('2ZL{':^'_߷෍TEVT;\q#߾Ƅ>bjr2!$"xC\ I}l)Q"[knrmU(cevֆ`13JF;xdY&T'P{EڙL<ЇlE=yk.i+0`Mq%3V[]ń ,2FǓo!ɞD | , ,ػ`Nz"2x *r g! ٠ L R1dFt0vyXHSrڕWY)!23T#κs7蛹Wg~y + ALW3$ e?I;Hiފ|LT[č :He݄ USGSҊu1"hM q.k Ȩc 'bB*M!1}](1)/< S8j80JMv.[Ih_֫w:o2I>$7@3 yP\E!13!ѐR&K>c{c4ٙ;е? 5p`$yq@(,78i$rcb't B2rU{w(x d 2%,AFЉ4Xn^tǷQ÷6?l;Pg5yݥaIKgFN+֩briEE")Me[TY؇޴_pb;gwY:݇m^/{Pȧ^"_Zo Qв_׫M` xӎ]^EcjJVj^Ͻ +N𯻌(h?=@6S|JieוZ{ܶgt=Z׿Q۟6ۛfxr218] pefi2I.`Q7.^˥g_w֝oò4V-KA,H{Y󂍌qKOS\S`=?gZDԹ;c]'zgt 1_ks+r< 愳ML)f>9VtM7QlC*=~<v,߬XE[2 .BREoWE7鍅ty $8 ҍv(ګƦ^l_o[UUGmIy$@ +M>?MB^b\RT`zA #on|ctb+f /;x O0PJv ~S16ķ{'!~ϘݱfaN_&.%Gp 5nH]"U#ܳ?-k]sQU:,]/yyhN/ZK\l2K.۟jeްqQZs/3/H{xy 16{P-bz*g7ӽT ^M$8mҶjim"!1൒[oWoKlzd]oa4އ_B鼫~ Ueu;=̹.-{KPD!YHLo&e B/O[WڱxϿ~d9. ݊dlƭE1幢%ph9{ysoO꼟'Tӡ#nE&6_?629:,tҡCJKLf48#8==!T M'_U;R" n/P7zP; [/$ڤ?Xuէŧq/.@#w!TPk\:K0 F5 QT'('b$j xr7~ h:,D[& ̿b\x) E@!K08\RG*q1aV67|F`\p[2l]S%fFDAu}vNЦ&Col2׊+y~H{Ci؎&r]W_7bsB)2H ΕZ:=0-}ف]1{ -Ry8Tm#$.TWD~g"(:OY`g<`$.sh?0nldd~]{ vĖg<yy@O5eX͟(#7f^.1#4VOZ}`23+% dܣRaJKnD$!¢XnyƪZ-tyF?꠩E~X@w3gtlҺ%Ҭ6>C#㣚WQM\,v0""l%u)Է}6/|UJu7~I2@LҌ߾Q~p̫~"7>͟O !1Vx6ꈻJ5Co,-rD2I:{>J,si6+9NJPix`Qd eJ+gI19O3?]j~QoiwH b.@k__m= %> G q4?@`p!K1}N&mDHg6݆ z/x-iַ-,|tb׃s++jO#)h;yܧ( `Hk9QW!'P}z+KkY*8ʝ7z*j|sWLc*h'k׸nIw3Vrũf͞-d&>9x,[u'w$nPo[]שhX^ Oz!)z4+*XѰ`oi w)Pʍ"6ՑaYqGkZ mD#y4-zjP/ᩪ|{@Sq~9m&6kSr[q4ۈ @Q@DB93)e]SͤĨ\p~.%·Τ|*V3l?\A1%cBvOUY?/JkmH]`GC@p:%@r b b g(߯zIԌ8C8lzu}ߏz85Nd(yv5cOY\>yY}1Oд>az|֙RhY+^`b壑q- c쌳JPd\jwnkr5\s75?Fq)C"P]fNyaS b&X0Nn6Ց _cȪZEO׹!ʓH&H -q8k2= C> w.u)#,AcH"9OMVkkdZeh$ۿz3hFH%#Ad#J@UR1'E E19mǠZGZ |[ H㒟.͟\OY62_x`Ñ[$oKAyb%y\!=ġٯ/]TO-mˣl)Y; ܁pD# 5:{p1ug(•}:+JPWgag?@0[i+,krTar%V Cn~[]G!1oal|{MIצR%s*e\kyyuv!0ZR4BH?.Fb^q?pV\RXR`Of1 2ꃯV baG3R}Gٕ@liY%z^]|IQȿdic:Sjm]ӐӬiVkCq$O Bws得SzufZ뻝TTXHHv<.M ^qr Ƴ A%B3jǯra_]2*ˑ+w|a:&/~d4O97bnMxϫ ة^q>%7W}wwツojzk -ᗃ 755̷*Vgj[ ^c^I|~9/n+_̀]qNӭ>G>^iݖhVJشْ1c`P5hR\r Ti+19y4H)ʝ;S\0_ VSSG:a9QmiD&2EQX3Rp2#wYh¬?SD͵o,_=2 xE1:rb&da]`ùU.EogD'ZIB^}w[M#퉵Yn.όvĒx,!vKTu{ : 6ɒ=N JgB21iX" $mh9ɫVm=J:om =?A<9*@dScPck.2[`^% e;X6E2b0L{|lg # 0%V$s5F  ?hlьGgQuўulqL8VU}#>cUu}G'1]s?478~Žs[x.xUkړ$ǣyX\ ti-erO`ޑ]RAXtIaŏ^=44mRc$ RӋҵrLHr7yO ?Z/.?Wo :RckytXFdVuDC ;=B;8#QN,n~Ku\խ *#+u(Sе.2֡!6#P̀.p(3v4,xҸ#\ | -#d1q.uS;b"'1,@uժJ\ Hory''q$r#z Y.dx]&r5 NZ[ ]% p-}UnD1vBuUK>эCeSa~0;w'{CWn9zʕ *j}'C|A6-ݼ&s.]U=6 ]KHC?[! {qOq2LN'^%/CGU3x3T0mMpʚ#ւ|;(B7 JE;/FfJuæh_lc!=]@j?H'C|I|Yq^m}Ϝ{n)rD yo(JzoϺ:Pn?l94N'rbn TXBFbǜ*3Gzik/"kNqrsֶcO,|r.|5ă9e"ro Ϲ3pДu){s,&`ZVD0lQ<ج*HGk;ҡZ9AE:Xʦ#YH6diI__ndtO4t6Z7kK/ Z1\@s܌Rmعd`^4n? ͡g>t,-((!.?}HgZ2zw /:_bTQ:Mu+n=C{1ȑX56xa1S(7f474/i8ƿ1>Rao (6!4o %S^zm'W0U &C>2tepÓZQlK{ɱR=i26AfjQćR5 V,YsެvYoM+U#}iay9敩9죷@;.DW<ŋgŶp=VSNWlt Y0R~tQ b;].bTn3q|sKJߦ#_Y16 y^|Ps8LJ8 \9Mug5oEZ_[ܸ}Ņspwk@bV^|xQ4+7;Mn.x< oS5?{-ب&q0vmXn_cƨcrjD~w6֘r?l TQٙBBe,YJL&0"4)u)X}T c&:7{/a XiBJFӁa-2 =E$[c"{&CD{f*K)8%0xr<!8~6  vs!%;N@$dѴj7zi}e=*zv2.`c,/"^F@1Uj[o/AvKw>*) gr2hzYb<-m<072"3#Kdl/aN(E3fͬ$CZejj}lFͫ1{AVP+4KZP M?+WNJ'|JqTQ( Ȣ`;X+\$JM@C 6Z[ÆѦzN#Dvc9@RvsFX?y"Gc-ϖP-}ϖ>[ޛTM*f#Ei k, 0tca7 >؃({zׯ{Ꞿ{Ꞿ{^B<09v,\tɔƠ5jN+YSZfPaƘ@> ڵ,KbDlee ƸKgbc//L{CVe) nZ1K].`Jy>e%Do ߀=Ip Rm_I/;MduE^% k$>O$ЮPiN87fvK%\ZӞ4TWf׸XԹ숬2O?7şsPV*?OE|6͇eۺmcSV`1uĀ*oqSKof 7yĿ9 *DqLmrK*Z&w$|!?x,3]dO㶫fz>Y'%K}L#v^Q&Z854]9BlH~A_GPZv_ b#A`㔥panvp>,G>ecIl_$ta{}?n HkmHؗ>@͞H,}% TSL[$KoC%[b٬n]]Us3O?GTD+vk7u*vFquPA8"i+>u3v2TͲLb$' oIJ/|F} uZ1Tg9|}xޒxC z | \v͍}=vG}F]HJ:#s^ #J=37yRPy3>:^WY|jê|ͪC`t'ꩣ^R=\]ȁJgO6i Vclrj4aƍAT[Wi k/UR6΋KM"HNR rz:jF,B0#Rbe}0JN,GFQ4`pH~pQ?[ {^++ lw]-[Omn/+Sh# >N hEz\>.KGdtgYmՃAq1FwS I+;IԃDjjhG]z1={F\ tp[qCu$%T~ocS- Lt)T⣹eF*}Ygl{wۜNkG*mNJSl_*ϑ {9 vD8]E|Ba"D3oHL"3!H"9bV !1BδG.3C!0?}Nƌ3:QAsih@Ⲻ@[Χ S2(hO!)~!5L#(嵷8E;]HI+PR2pDBJAY&D43ˌ3YOgNxU(Vm)bL(b՗;A Y ea-4ёAf[WaisZ/3D-ݧK oT#N4e rl]#efS В1rIKD$2 RjDs$rŐ*"i8l:{fo G^ @m$5%h ) aC.!?GGŨOL(O9y.b7txg0@gT:0UT-W{;V}mmK /moai͍FA:#mgRv梕AѶѺUhbvGK"Qjuqgs ,3kIU>bZLcʧ|ZN̍Җt $d$}I2qG[wttF!Cu5:AI:3TQAl |^\Ec;&zvikJH: 6y`YeW&~$l _\ Dϔw,(qr)͹(fON''|&5GwP{%xz5B)7j<[ߴ Ȣ`;)"VN"xm,x…☘HrnGEo%S/'Y2іuW*HTs_q"Y`N(6y\挢\KP(^| e)TGS"e&#QHYu2LwZ D佖w&Z܊Y2tqRofQVW%#0gx{~!E([޹G;RܫbϤVvGו$޿(luvŝMrԖIS ?.&=o{: Dj/[wd|b6ݶk=6" NsvzHUGNR4?vV:qU6É5znƊ4ۯk~w([WnN ogTu1'"0f 3ve$ c\,2Ml: {o>kť'lA}*ޝpHp1Ok ְ=)gAYټ,.d/; /j_wy!4jʓIZپ/:EPoժAǁ! b <* s]`HZʩR f(%R :nc/;Uew<:V⼻B})nG `+8Q[M*Ffd j0D@a'X{-¡Gi<{u n9+sqpxѾǣEe\`ǔlZK(I| yIEt$LEt!"'U`6cngBXml>M>SZ.z3 (z^B&8[ӫ( tQ0{ctu -9v"g`6Dx5c}wYoPep09c^z| )V0g-F.2FZڒ1.yEj&5Q0!S&aJ;f2Rʃp-A{^jFuԨE甡S~yЧrGs Ay)C"P*3Rz:f)N2KO%Ăh`l@%n|H*4b:_y:")TD2'TifQiƝVc   brcIP@af4Zi/u򗠨)dF"\B,Z-yf%7Sc}|9:m,n, B CQq8q.9XH頔]=;)# {&h}dSR~4=r!!3{=H+K>PdO41쯱Z>F ;O^EuK}>rz?MDɦQ>d]pV`$w+"aXg3tv!3X+;9[Xյ(85!rTat%Sj5] 2rrar1Lt&EZ_9a| 7̔Z%XowU%/:?WA\kI1 {tckXՍ^L*U ;wTΖ͓ޖ,|_^xs]\Vm*g d̗$G˕ve" OEv&2Fu$7t6 iAi;iɣxit2zq;jTl9[G%hMzZ[JZ]P#$ ́|5WlV[TNK5VO_L?G ŠnꜺOZ$`+(JD<񙯿MW udxk[n5mywyrFK7|7?.?|ݻ!&jox.Ep ڦSQ ?ohƫ3c\Iqu\,-+_aj@_0{HHwug#3lIʧ1f%>Tva^+nrl}mc` *0wHsIM꬟0uuPêwQ-I;9Df( kh#J,3z#5'|ϼKe.eCqH*"x*Ӂ~MgU3x%eيz%)t (ܠO?HQKOF6- |b1:ѸKpU2yab<9⎰vWd5<_]l bR9c6([8S̱u %}60cL ?.,h7T^% 55^1DC")K:I4jc::p&H$&hVR10x/; 9-r͙A6ޕq$ٿB[#"bw1BeHI .""i% Hr7{Y2$*2.^x^/_Ư 0b!zFLI:g/ܟ{gkj 0{v6͹^BOq%]{cx-v\C4oݑ"G-=^{$g}$~0Sv},n>> _ ZOVK4KǦV,Q1J+F|y#}.}t_s86Tr'B[+L~%FRWAϬ؟o/":5Lٹi,٫ s~}o~mnyNg>шpnܪ3\o7@ ҆v nAfӛ>Mzsq-j֤ꪭf 7cH^Rl@ڏJ߂k _~wΊKWz|WO14*p"j8~-_*y(J@k?wgǗĥCs 5[E;&i]mJELW@ $  |%f&]Mw}\Θ_]zͲ]<Җޡﴓa;Va;V;Sd9)5_%+QO`Y3X[?STjVT!hɸlա(Ui*.6m|m 2`#2TWG6_%OzsiչܖJޟ]ƕS}~{*ܠ5DWc߫ZP#g tj!Ъh ZVښc{=c;_䃔=){9 g6 MNHz&:Z5:c9 ؔR 6/f-foE:<͟y?4MjBlX<DW΄#[dMC:6ZndDӉh)6MNg[Q!w /Kf̷ ԫG*aR֡Ig%1RRDZ4H`N8kZyfƁZTtXCeR6Aze&mmcKS^WUScO(ƎD1b+l=Co$~~"@^=yz OgI7wd>Pף>֡s~wDWa5BO+rk(<睼/[G^;R}>?{ +?pk7[`wMO3|f=L5O !nzQ{xq:*b)s|W&h~?= k.Ґ)fIo<=V=59xh){©féOX9g!ŋ6{xaj r*{ۊG/?'C|"-w=Ox u~]w~Z4}`g_oy=uκ!w#N:^ǝC둚;Ӱ1Uweyr޸ @ã},ԆNjϕ[37N=%]ӓ3@v~#wP6/&GMwA?':(@÷#[R_ItG_S=P2 ̷[|s&Zs/5/ܺQ0nv"g{g}5;xz(nƖz^Y|BkTTYy\wX8|p7y-[#/Rp3i^9 \ XPh!C[u1IQUo״Ng.ktFfk>mY~ ?x}W]3O/J^zwKxj.3MtV*JZ򰛼JQؚu\b--iF._Pt%h$ҡZUbK%KKQm {I\ݵhj^'C:ؚEVn=5r)kS:,ގHSIjɸf!,brKZkZ&%J͈j(7m[ҒD+ɸYkhK9{~jΗZ'&coVi\afz\5IUňkYz _Ʉ،q 8QDG2eUCl9(w{pGk83E2_Zf.Y1t*lDʒ&Sʈ)5*dȿ%`w!\qUl|˱yg5; ߨP[w>:x! 7B"1izUA6㜝H1dL.D1jw;3yYt3E5(%YtV>!ɵs|nZu>Òp 6k#%yIchcj,I;ml%5$rUy4KZ!rE,2f;xrRfU.XEWV+LΠۥjPKEu)I7 SI%2] 360HȔKK>(<7+;8 5`t[G`CQY QJ0P|]r/%, Z, )Pq^ռ_ ` Ʀ pv@P L%JE5nXEWG>? N?B]R WE]%Iw_\08P,"/ՐM%@H/%YYpP(k4V Peq>uʰ&TQ]k?VZ}p.idȀg9nc*&**Fu(HQb*whpD9!"" uqX㾷dvp31 zO'w#ҷ^f&&F9,.|݈Zu~ .rY*#mFAv0$}5)1J3A>rsco٪0l2#+O[QW TK\ \uD߁Q > n$?0Ӣ aT:w zaE- `CL.U 6csZ[͈ܤh' h*U|lZLO\5gBr ڦ5Y:ThM_t'f$*awXUɚ*  R9}ب-,f@h\)?RLčr,Q'(\zݢ>+ERUѴ]b7_9eTǤ<Ҭs2~$!XO]Xv颀+Y-0 U~y,_ຎ GFH&ՀߴzFy/n:88- \: ׯ۶#|4y'wrZEeeʩV² Uב1jky>q*;J Tc]d~sR9D{>J K(`-'roi(E%POh(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(h(]%P(I Q\|xڇV:JoQ {P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %P %зg5>J (`O^ +JoP wmK/fOv A1>%)RCvoER8:h{kWU@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ T@ |@WvHH d4i Jr(H V駎B* y"8,#                        nSz+Fj>⬛\?I;q}' =@ Hp pK͡Z&

j8 cgA[}1?ZA/O~Q^T5*j4ƼMrBgeS^^O,M&GCo_ȜDED *3#lˈiDNG :ARK‘^2 VJ RHye`]^Y,Ҳ7)9ESf͗`C$@UY-178.WՂ?6eLayنo{sr/2nhRx :Fkc:r}%cL1Âe 93-:4_7 A+|P²r&Q#S!`YʚPa\p*ʹ ͘XIEG ffR/x,~Ȯ;{Hnɣ|;Wfn"+'Y2̘sy[o 7w\fkN"9jpr|,*Z^{%d$@Mb=o$)oNHsgD6MVf?^191 }drP(L8Uh.b=R=W .lzwg /j_3A"^ YGJqx1'W;\PIkǏn/7< Ύ׼k#+X|gC IJADzȘ"H(CLRJyư@I[tn'Aܮnis-Ԫ,ʺۈ9)QBנX_X,X2zXBܚ^D,ŬEE1߰HC-M(Ҭ*CNiF`I)X`ƒC7 +'"n,Y`"Hݾýĸz5'dC|$s ="KkNun/7[SR:Kcr&%ؕwԔ܅;2ri~A,%D,X |v# %9:#W>#0Gggӿ" .K@@I˜30OqQRTq私)N hE?ϝKט NآZ)zQ&xtUg&6L4|?w]ڌ&wwƷc-1ٮm/F MbQs>u=;s9g6@?_~ۧ?cm/MѸORb퓄\ua ˏY71"q%{^tG&gڧ2I{ڹZkTp5LS q{n0lW5Xwvp+ՄA'g3{ W^npⲗ&ܒ '[< 3$k ʶ"UnI}r x c9Ej <[fl^:l\+vmT*g~:>kʵt64bwHZgء;tXG8ጵǍm5).}dك?`i |4VZT]Ӣ̬?,nORRƹt["Rz43+!d`$POUu{ m]<xq"{v۷hՂvTVSQTV,J/#u5O|jn"[rgхڐUxA#1#'ZfƃF$-O(pnQgcavmt@4XӖk  S2H$Fh.]klF1)9u'R=|vB24SK32pE9Xri]HX~{o ESP˹e*4,(N$Q')74Xý"NK#\Ri2\v)cPۖ`ې@wmm~yaa ,0R}Hv&~-YVtLYC nv7ͪ*x C]B\\ K20{+ Ty^-WiXL҈d0Tի2ԶHFP@GQedl]Zy.^y!hk^fbP]p'~`,g__L/P!"A[5lh )j&Ix8[(RZo&ޅf2xˣRhI@\0P+g8P 8-a@MB7߾rkDsW!6kP\%i#y鑼2yU`)U: gf85c2(PR ER1|+i]ܫ{p~);S&QS(f+ QD\YkR1Is"SB±(M)M(M(MRV$*zȼJdPsD=6D=q1P4ArR 29X[B,c@^8974R!4SaEbXѾ\%X[}!'ak/W(ŧI(a6Iac4:MD)N 'ϽG*T@*$AT3'oN'xzL$:5Dr%gN@*蚘Z!TrHr\3%S@G4aI0 hm W!x!&mΎr՟o@ډ:4Qq 1R*ܖLEc9I 9PhDԂ UJ I1{E>6{ȥԑDMrqTDC8G`QƁkF J**IOzA:svQm㥭O`BkT8QϾBT.8BUEnnNDW9WyB%߱#]̴b'F6b}rC܁o43rsAѲ ڝk}ggx7ۮ6氱6X?ignʦ|{:a(0 M-0=An, x͚"ҥg!6F)rbUeVfleVfleVfleVfleVfleVfo>2c+3,@v_%Vrj%Vrj%U]%Vrj*RɩZC*9S`VJNBQɩZɩZ\%Vrj%Vrj%Vrj%VrjEQZɩȩ&:%s!qЉquJ*b$-xVqC\EٷnԾV"sQY_P7c]U*qPÄڣNx=XRHz=WFhkL!k4%$a :'>s=sb ՗ RL7.<(I86f$!D+$3r-c1FH+D=ˆ@ / 3:(xFP%gz0b )ͩj(3ug;:d(sm [XFrksGB6e^¥84I5~#_s|jgr\}l7څ$RH1&'^XП{rZJe[Nr7r rvTBN2#tɭI:zG"#4P&"3Gq(DjLi’`B`IAL2(gya9+֝嬕?߀Vi7LCxI#AB)mIuM3O@aHS FD-(PEܟW;#LRH&h!Qatu(5p#%jZvX'ͅ7'N0E*}Ri}5"o3\qD8(ڼ+6'ڼCJn,b<1X59)9W\yBNScA{BcB*\ICr+"ط 49Z?TWtaӽ "o ^FPD)Qf|a8AM)V!9ptX `XP<` 㛯"fȾa KAٰ7D@o @iD;$`cpM<ؤg n;Gh&~x>sZgκG<.P\:l;q*]_p|?ӳzk:logK (mسZR?<ڈ]4}ph၌~x\* >(Y߶x χYFB<7 zm^CZ|1n?ixwq5=`fg7in2v=>`_/n ]6snW'?}F}`e6╻E yp 鱂XSDs;qt6jIMbIl ˄5=smڐsQzrn.(ތֳ\^5'М%Gcѻ"=hRDɨh:s'y4*7m׸5} KwX_.a4vA  An^pݷz42Z_gO#rd'n"U2<qa>Cm`FhJx$XjdqHnϻ#Ll ҁqRg1քz_w˟\.v:Rl{eeVc h QxF;:u ,UV8%3xC-@rR!Db:Q/%,wT+Q;;t Uڛ[:|m⚵.Nj|olNԱU&;pK.+NeA{ %{3ֽkw%֬9Kص.[]s2ڞ|La7oHmN4˭[7]낙vKv4uloixCfw^hy>L ݞ;(]hҫ-mԜۛ*~7דJde'Cv=7}`et@tAӆ~kKo(i`'" .=Q!8.L̲_ym63&4NgF3|=&; ~kkFU8,sH;6GL DN0S2z&x^rf! j 19)ˉ XBH6L"8hQn4#Z /'q9PiMdy͓מJi§ʈEr<cP?8e `zDpٻ6$ .#[q:MpI.5~JeRKR߯zHJԃ,$R'~hfȮǯr2h/%/l?b^>5yPsvVckR7ŝ~t=4dp:m@k kAҪm()0.3%'dW3Ұ| %0fg*AxԙXRѨl4kY ΂ID䙼Bîzu阳RC5#h /t"P*RH!Z&I:* IBT!B:&kZi==Rtxl';v=FHaWEɓ?vՋiFAF$ 6X:w/wNt=f#t^)~|=)Z/P0C0ɕdF?<2QJkhcAm*FHR"i f`]୷Ak&eۇDND UuEH6U]/hkgNt&G^ Y|b)D)}6 J:ǐ_uēiۿbP'+âMLVDK R![)+FgIJbȞ!Ux=[Ac+P,GyTnol*]ğPQCH($($ftAe@LsSv!H (e/O"ӏKc(ȔK(m!U% T6Fxu"-Y@8Z RQԭ"ݫ hW6xmj녍9+1{m9%lxE-jDl0Kd%" ||W0=5 3sN}*%k6U򴀷+2i'-/jxJ()$;SwDFk㞟Npqo:c3~/fqo5"{`*9\s"EA6L.䦏̐ah~{Ǭ$7_}oo~͡T__[2˱V [GIu#Po@ꧻmDWCxˡ M]z;K)oWǕF]Y|ZbA qSrϭ[HTg_lL5RRT;~i5˝%}ym ?\1U9g1 :$ɩ'~zx|WKukJZZcP@^$kJϐALR)EcGGp ;f.{2ly"). fv&y^{`cΦX-}`Eq&T'P[պ3x&mFL筞U _`5]lf]İ4[lB ,fREn6)rN^g3F5[7YWC! Bu.$J^tJ"Pd`. / ,[ƆDN `Z7MQPzKoii4cUٔ5QRI9o@H\Q 3St}$ b^@12HgE@[ag|n.YPa iS =6rl_Yr Y:!AL&Ɵ0_;CtQMA5Ds{{쨛{O3^drD¢DшUફVhn~jnx3ez:1 /1[>xl^_hjhz2`I%4]lFhBI#藤LȖ qQ* Tha|TAAŹliv$PL"^N.jA`\!ecr)lMb겷:Kz_x3r18 >6j>||&'Bx: ~DUġg|cjxJ܄ql^DeHT yS$d֑W$)@'gɇRN͍Nb lBNJ㜗>k6;zWBبl^)hJUkpg|ZoY@'K?麭ky>D0 Է,p*Df94 Cldۆ)F% m.۳-_Dwb >ƚL%8>aQ:BAT LZ/ m(K4ճ5NFF%l@A/Y9)5슜;]9 -#mʺ^@T艅4^g-)QR)*ɓ$s_!H)ۣ)iTHa'ңVFgu0dІsTԠ^- V,< QcTkYV]T/N &VఃeI: /`eВlF#BalbkkUmGP+e&E z"y%ྃcyյA}8Hp)iFaW`fI;ʳCrH C WD xg^|F |IQPY=\O|Cڅ'@|V{Y-dVnםmicU;h\TP5v3* /0HRv\UrsUմR7W_V(;d`*v\Uj \U*W+1W bee{h.G,,J ^,~ۅyW S|<@} ۫olhbnj?/>T /r\|vjk; rWtݶJ%L?3Vłl6<}o_`G)?f#3jlXt!{wkw:Bg[UowSiK"g*:kỗwڞ~q1d|ˏB8:;᷽?i.-FN+K VTŪMۄalµz6|ga( >EHT|ҋ`L05tOVjLy9lK%뎗?F͊Lq* Έ15@4n,hޅ"}@dJ ̰l^I,1hOEȿ)IhRɰyd{ ԻX<8/rQm,,H9bL!<:D .:($! > m4|0kq.6J'KN-S.j%Ag䓓> IXߐ]ŧbaʒS%RY3uHg"[PM/ɲ*o2hY䳂.Dּw2KP.D"I$4 a,7JНg/0Hnʔ@ 8A `mjlA42ʔЖ)q[WnQۗ)9UG6Eaepۤ"9xeuT޳t Ϝ |RnK/P؆dfр)؅l)v> $zwjC)/ueX>+9nQeNu7Qu.<cTKњ`kqxK!~xB4V䨋^s܆Qͤ&<; l3TRwem$I~`J0`,E"|X'3 _ x0RiTQhʺEI9ph-X+  Kf3VT#t|X!6'`LeL}R;IOڐDz*IK`tyb%[k\NJ>3^z+ӏV@ozOMM>'K^9F)V2"""A #(H8|DOV5&]!G2f}},!OqK!ƟҜ7L;^Cw&/RK9]M'7')7{%ykYn\lYz) ۛ`H(@Tšs!"NwWL;$o\ٙr.tN?go_'[$!K  Kp&tIϱɂY>AI1y0eW.*A}EOSnۍ._nN镫~&W^wT 0h7%hKs|{Iej =rnoi)ۡ_t}TS*N@O(1|hr]))DknUw}a}+??燷/uo.> OLKXHw"t,~~ir- KSŶYZf÷XWҜbݿ/c3w+x*/czF@v;-lə*F10Jz4)c댾Yc~؟#D?+à z#(wRLOLyh4P_1~A ԰k(喤^C8Df( k#J,3躠ّV`MЉׁ!tzQYR)Y5Y6Z s\"*vV(HDC,N+OnOy;_-tZvVǩJxVoHyQ3$/^F98՗>U]Ny+ޔn"lE%\& uC()t7r?Q~myvR)EE@(KlpU+9=rfvo*y.zצިdrЂ6Jµe _7zlm?Mm_`5Mϱ ׽_뷪OLwuse.~Q޸ Le|_wԫF{գF͎{HM=)x;|KN3m!h➆>!p3[EJHꊩpO6·Gb$i..׿ MtTQe( Z:ννi^l! AHRVa S X1c2b/Dk45[!-llY+UX۪>rE_ i&Wv3j S {׷pf #Za`n,zLNpcf7LX՜ -3@ZjD]16ڌVIg@s1xQc[04Ggun>AݸӮ3 <7r=?LF5W7C6|Ϗq"GG7LMO7])lf{m Wmś?>vc7![)6g!:d4@CeAOW*R@L" w!{o!Sw8dsڳs ܛZ-\prvXxl,|<t{E+4"{r{ː<-Wb5(|~w]刚AYlN^&*o [NZjlGc=*o-#CR&QCEE`.0 r# T)BDeEf -be[=S/Sj%U+o>X# 6s&V-Y%05{ީE˽*JFHqF@H1rKP5cdF/cL%[3f#gfd}B:n}&}A+zESkjӗo~=0ˠ_N_RHhd@a g'ZK(IZ @ %^F^ґ0μD*GxAK^C" B0Ah L> 'H !"p#A19$AʯN^iw4pFR/IN<4(y%LR#1s:CV$;I$}Jڳ=Ejgfu# r#p|Vi_`Kl0WjI9- -AƒEiC鮖d}(}h}0ae ka(p\ Ղ"yBWvEIr\f9BN `E=-xN~_zv}H*ˍ`j.wRfgKl)luN2sS;#}-E?^](]>17VJ u*F4 %aH2hu(joFI*SܣFj*5@yA)ıH1W\.3t٢i:0 D}#  ̐ QD( gI6+(}xg@}ʵ啳$80u)qM Aʖ5RK"(Q6G~wX#HH#4.teګ]}W(p(kJ5X$IHw TaɩQD%GZDC]%u.uXERK"`3!Q25 D-TyL(LW]5ä1bQ#Xt`3L*e)\Ǵk300/plsVAQQ/p˗Ӭ+Lس~9]W3߯ч]`&~2Uve;&YZes6,:CZd 806*Jr2q, N+<\fq4ȆJsNG@}7,|ύƻA-@}1ۀ0U[>Er \<˙ (HIKlEhQ$(PsO[N 0[JüB#  Ary22B=Uy)QGu)>R'8[ePОXBz'*%aE1F)EZ;E;hiLN()ya`"ZA۠FH"I`ci9I1]۷Ei9ࡇ[ҏǹB(z|%o+Q@RKLL0ͥ 1δJ:%@1p˴RkFr. F;mrsjF)8뛣ΎxZ[72 N4e rl]#Joa C3* 60 r Ȉ&*HΑBJ(2Lg%Ut Qg^Jvh9=)H9":%)(D#g@ Ra^2tHM: ho|k)#/Q* +oi 4 Lh2PN:U5UˍI_ǧM,u2~Hkn<6 i80 c02';1"'¬Ѳ uhGA_\V+[o;oր_2OˉQR0?0ݬ T9iUf"2tމ6qFB|COL|d[s*{kx垁 LQ֛tEنxbީǣAl1E{ɱ6$Qs-9KÃ&.OƷ ƥPVz|30.llc޳[x%^y̒ʮJD*Q;^\ UZU/R)}Qa+ :WQEڠ+ .fF]6;BV($?j\W+zҎ҉;([U>~ Pn?؋yUBYr0:-ҧﶅK9\!r -%TD$@bv\:yc<(p! pT`lJDPL8ZBPH)c s]DU,YydcJD2p8%P)b|$S -fT5R֢zjz&Ks2m,s)}/;9nI>}受|B}̉`"䓊b.+(PB=Ƿ@cw̦Vx7(1.7.h wK&n|qO-3 c+bB' ^ +6vpJf'YӸA>%xJEfw5%>+YM fwRCMT|P{df塞dSzsrhS m0gBJ~/mz][%9^MG7I7{;$mϬ8vTqUηH0毸@Ԉ֛{fE6.GĖ̦ArE?f.Z@M=+3h#ަRhʪ$z@Sq XuqoiWU 6Nz5vV{<7.i7EݍF<|ۆP)H(iRZECpk⃠ւ*d|ފ;{TcnJnO0n3鬿y跨6G/V!]Da'rFm&< Eyh|Ɋ:;KSf`e_Nj_/.ZKF(E6`i&FMl~8RhRuz\![Q:cF:[ūyp:F00f\/"gR}sEh/qU%ChH)%^6Ūaa0ЇbX(XIt˛1)rT֏:QWUk}Ie\S Is@\zR~;ΆN%BH:[J?dL*~:OG7d.)|`:ik\2ޚxT`z~9ɟ~O_cgJMio 6$d6"xǖI#AWn%m,."UoPo^}¿L4ǫH7"᧍|n}W?"]}74Ӻ^7 o]ͻRb/B`u*}JRJc6D72*)d;T5tq F_@ voVm};G VF<12.6#;~px%RSǑ8FϹt8Dw m. Ph%e:D'.ީ{Ց%,+UcVHu<DO.PJMLϕaXc|`+J w@bPlBUWg2֪кV׈oVRt{JƗRN(5V!Ri,@)`K N \ GDxu@t (H]bqHtŌ"TIB'miAz"y\q29rp}(Qp/LS{t#t5Ӧsthb^:軀@6hKMb( DxKiUb1iBE!3MN.;튖ѡ*4:g QD<:@6z@hm41)T(Stȥsi=76tyV2IPDRBHψ9CcIx4eg܍B,M!f,[sXzo>b]ࣇ;to*G?2]qTT=2RKI,M44Kzx^)^ J4D*o.0k^AJR:W ߠd:FI*D,N YíNiT3rZԶ)d 8Z>Sg\݉ |騭~t~aW嫒ETAp3t3 ozx9lr-eyn:lݝ Sx0Ns;K7 fױVo:my0״7{w.F[pEtijK53ќvhۈ&??E_z|79P .ϯN>1]) Z"2[C%V_#e] ǜ}p.Fh4fQ|xqpx}<Yy]>BsϬ@[x3\a_]dw%B򰵊?,e_.orØZ!V)Mל &8&xNrfbh#ʇd]eDO|v  8[-GPo45u,cWܭtT)wtsm΁t&'V zP;w" 1K1))fǨ.k&Vʠpps))iq8ehzRĄ.%au *n)&4R][ȹ2vR:iazٲ*Њ͖OcĦ/g~@iχƓ'`: $R1βJ{JL &1a+x ]17(iS]hXUYT S'fx5*S+x3k̙$uDmmvkɯױ^Ǟ9@8$tbyЂH.8""G+[Mg΋.tl"ziMM;h ۂRtMwK7.'U8 )4A{o23Ի9D3[qSQ#gbMj GϗGuˣx[;*!+b8lCԸw?=Q܌\j?~fr]jޛQ)<*NP/*HYJGk\Y5;^#g`j[28ق ;^ȳ͡2xo6^>Xl=Li0ޛ8Uv~K/)?iX5[Lcj4?Pz]QDz^I*m;>{3e F3R/Gg9v=诇=eSC]Nts98*@ -hi7dYJS8aзMAe;y|P6ȠwZe9JuR˨3dή9 HZ{EԈi)ZYfq/ˁ;Zh;ٮ ^6m/ uѧl80 N#\w/d"?,#IgPe eNG(,͡<6u%H'DN.F՘S/}2wT!^"$i!ulwv}݂g6b3 &)-?bU5)b{*P$HUh| \ 3DNzWlݏ4ض2lx\:yp7 &$cH;%p.&% nk1yPJp@\1q .Dp_.JhMC)<k1茜ǂGď(߃"Esݮ:[6U>㻌CڻW3͜ةs)X@* RZT F[4%tKُtPTc\XFz $Kh֌IEJخ8uEg==ni)ٽկg|E>E]a_;;rVTuYu;p!۸ okgʇlԺ m9{\d rIhĎ1eңs@]DǸ ָ IUXe$#1(ְxg٪\\ǫ./׵x{]~ƫM?&W :<hth`ZP1$Sy02`*ej% bJ% bϰ]MΫSgbl ?f qY$WmvEGapiGE sq06+XLni3ZZ#4Vɥi5\gp"!-<,@ '?6Ȓ; D &ǜcF1FE~cG{ɮȥlf~23}?*bq3-z3ࡧ@2 ,ɲhq's4ɟkĚrBu\"k>E1~wvy5[^S7+.3_X< 73pz2I+ %@02[cK—ZE(;" 1$$HVNQɅ3Q#\@ꄣВKTuHB=hN$nһ:G][o#DZ+B^_6PȤQL]2WoKC\FvW^]|w;-xDA\"2Du3X%.=g^Y6Z-|1zHFBqјl2{el$|DLug=c/IU̓1m76o$w ܭ·䖖YZ|h4Z6ytkm[jykpv,/XH|ݻoO߿KE?s~75aƿWWޣw?^]U-T]z>;ԋ.oVNx5&Fσ[g F[mXi;;c Q)gRT=Z"=~1`OsM :3Iiѩ?oܬE= :xԔf6A0l$'6QH:1Ic04G"t< (:N#r{NG CPd_.+`?. / 5c"L ImQPz_'JXE6e㋈1$(:G>G^*K2`UdcNѡ6F*D1&U@&*ebQ! *VB1:9C UN,PA/XhpЄЏ+vxY1Y UGS[_;Ųf2WmrG2+/=_w/zt=mQ8k;7rk&=un}Bۋ~nR1f׻ۋN)ty#A}4Άlub oo~yJ^s|[\>w]hXvʻM4#Moxq[<Ssŭz$/vq`s{KW/;-GDXI,ɾK2dMDz 0>neҞY"5$e٧E' `UMDyb&n5}#m5+ 4VcVKI~.q9M"\ =!L%_RP F++ lcMbdnAS5"_xR# i~md}WQ 6 PDAkŮqVSl<+cQ)v;}t=1x);ՌTBUŪ'Gd0nLJ9YiҀ/֌5q͸Ev63 uǺPΡc|I;On&q[~g74rr0~GH4I Ҫj4m1ZH\RSΓ+ \)qOD& ^|I +͐KFC $uFE*+\3 6I".$e/c}A2H֔T:6uBP RH!dƜu+@QB]؆^zI%퐒%Bѡ78Nxe~/ۼ"zwh)s u$q._Ct xjDF0$q/ ƽtb b|2()(85L l}pV L[Ze҂ EhE4 MI}-tJ bÿ́g[\'<κJ8RN`}} Ra]>q6ZN B;dתE饙v*F^Ѐ q&R#bU2xyͩ s:!jOXXL`i|@4E@etda.D^(L|QTN%[lلOSR1 I5 6ó)WO"FW9;mV w/ka84Y[O9}{'pk͞#"+ ޠm dt(k[?{H7d@l7=u7YBi \S *,YGpbzM^$(Gmb%aW[L22]45!$9B]RZg6s7n^|p\7kC3Swa[v;=;Գ u yb۠{ӟ8\;xQa6]lHhBIC1^%'9Qb cqQE AA\egAB&4 RsBMl"hI!exe kSuٓΒQEݐ:kΎGcs=}9}X#y~{=:on}{q.nr|Ӌ=+$/54!t[bkاV $G,ic+^TCNѭӭV!f'uK5xDOޕe(%R&ZYJkۮY_WnɳbyWO+-zboc][G,H}V=L㛷_/']Z;$4jdSRR˽ y R < 9Se9\TS3v)4kl"M.$`R5 uF.]musV0^•Mzyј*=a#/"KBP~a RvzAV!JFFIսK u:)h67`jXg'%\:,g Xa5Rȋ !*QTzE:*KA8 |Rn#1`fzc`)k]N)T Fn%ozò@:\'aa |ÎJH`'p(|^/.O> R.Yp)FKq& Gc,uq$aT3_r3:LHY$O`"EFׂ*RbrEV%O, *_, U@9E mc+B.mU|`g= >m]dɣq_ezĦ;pl&UŪbW,"':EN9[Y#W_1kutQ$@4*"oa&c8s$Z9Ј%&v"PR>yzЀL bD`9G" &RҙZL-?t~R')e֩2R%9Sc .INdp[岯F@(#mіH)"c)#\'UFxaW!M6˴d'FV y:)XmgusK,ds #4$WM4I[v=S/hjxc *,|-~!vZgl,9v> zoP%)C1C(aAcՊ™bXЅC uVo,biR?<<Ɛ'mBBlrdA)1Hu3წ3Z&A*bꝫ(Gui!&,ty|ʔ04ՠ RA:X*e#Tu8tK 5  nRڙT2 PJf%{6 3KBVRy 3Lv _`Bzg5է]/&MJ+-g0E}W.7Vn JJ ZFTDH(h&c%Z[KHTDT>U"qC*&hBPqP3P8`M0GA9KV#.7_ͱ}ԉ\?bM8FU%Ӱ ~_Dh>~QIBH\sRDƴT\e/a91xSP7t[Dxg h͞NZJdnU Eo6eXw3ط]8@!5+ W Z;×yJJ+-bNBg]! ˎP4[NZo }m'm`mͽLЬI~>_1AV{J3lY9+.87kjـM_;eM:Q:鬯Z_igl 6*3naw[(pR 6A-:XĩRB%+(o$HsE 5kbC^$=sOa1xԼn!.h6:ʺKBιzݥ(<#O~,U lPLU gP U3+Fe'.j=xU݄5_)'wNħ RoӒ#wܸPym[Em1P8q"mYm QȻ M JC"g9SbCOW'q*q#ڲl@.J \jK6-ʜce sh *?Zf0"63%#S xBiWILYF7ZQP"EA@ (!7;GAcPVq)p ֗+۰Y.\BN8 G V1Ϊhs"'3vo}-M)(5CZs?6ߦw57\.]4#:psb.a/3.ц?_Ov8AI# xxUð,p00Q8yJ>nډo_'N2~!WY1TͅJH}>u+rDhpS+;m뛦[?~a_;|51$+Rxzh|9=?!ݚtͣJ>b8 ?g??ߝ}3ۿ={;\ e}_E`O&' ?ВkZl3*Գ b\K>q<.ehTF qfy7kݭZGz =}Bв͖jRN2lw[chO٪?'~<12/+c[_4] )^Sq h5z1x"@:(hj*;nёVaّHTĤ\jj[c|p(߷w$6lBUW1F>f+[zo [&JQл+.*u+I}\*Dպ@墶8eiWᰵ\R+"h=*h(:]AEEAbF*dI;+w`r>ZRH#+Pp{[Sע",q|>gEcg:{?E0OVl&MAhԌ"1T4+I*RgPwrtp﬐w(Z[Z*SȆ6ϕ ,6z\P(c $cAZ2znl $*$L"I!3⩣FksNʒT4e1r6[;çB˰rоM//~>6#o|t&LLÈQM>~lJqf)Jp+<syٙyyDT9*g2'8WtJZqT,e:T6F i&HuP!"e`T_Mn 8bl|0߂a9o3S9;n6N2@?s7`MqPYDM-w=4}ΰҍ}KD%ml66?} }6z/ b;n7r6 nnM~o7m{^hƣњڻ g|Ϗi :qrzs^zM5on!h[p2_:|L;d}s|@$ޱonef\u~;B4ZU"2[CV޹P}-ù[`UAl17狈ukXe`v@KH@[x\ ^^C,*ĊMd?2<#)Eh8Ah(#ey5`#gͺv= 3 /49x!/".].{I]kui0c:ikSTf^~^jMM;=kc/ksz<5{ˏ1i]^Sq.q:!7nF9Wѧ~23=J`7Uo+TOW\/TWqGmnitW2I {*ñ'pmBU)%.f&d$Ē0`GaN8 kk3`а{q `J%nkw5[|9?AiWiVԈ}h3kguCwn/>p]zv'ic?Q/hlQ. uY]U (ħJ(7qۇߟ]v[h<-: 6т4BLmu&R>Pw=se2>(,ߎqGOe͵A=x<;U̷sۓ h zlKLO7N;ZVҍQ rE_UTNP!7mBZq٬cތ.sI`kD^&Bt 6L]1$a:gJCH 9V[1ǽ3{vcNuhwXcuY7ż[ݶm ?zg2'}u3!u(_I/p,6u # o%o< `a vAgdk7Ld5p>&8Y PFAT%7>&G&Z BgZuLfVNCF΁[3Bs3٣L$fvmEn/vSww\55&u{K,"pp9sR)\pc 8_s>GN*V7r=\[KIVg>qM&'PAKGʁIُ'72'weVl],2k_kRZ94 hRV4Lpr6 .\H ){BB\H Ɍ̕xA>$BfI(v22Lm4N$u P8N]*K&9k^HCC,E-se9 Е\ Ǚ7kpzY>qeDrs6 k>bf}fMm UncA~G2lk@kјrcMX8Wyۻ>G}GiRǒ+3O0bF1dKzfmFzbL֭FADrGS) :{mde "O&J7iV#Qa$CT8f~DyGLsb=*I;.՗O(Ց2GNФL~ ✔e{MCa)lՀzmȪ7gdxzғBXtX6Ld̉y=H53e\cBU9D!Yɜ9qe,2AXaeR J93٤Fk$sVVY5rY_of-Nv1*K˂#; iΉt!c,9&C'&48i@M|+_ [vRYDV咨&:2"DžL4 $*߷OWAmqLryjfr+d:z1!fHX"8iq{ӗAxAÛ&w 4oo2 Fj,+gtuuZjO,`33٫2ϘՋmD?5{3 > uxcJT*9hƇF20(QsA[#a1IÊcMhͳS=AoE7HKvIwk qNHMΔlbv^v+"K?ohqb#*oHbcQAH\:`eKiymRBuC3TЩA\ TסJ<&fVI{2Cͷٯ}S*edvc6 $'/k4gdULRZ'E8Gw_r=S57nr/62O֓zކGo#][av}X۝ubڦ{BaV?4;{ْen7' I? B-xmNvX|@cN}Ut":ڬu._= dy>`>v㙁5Aq0+>T{P;E#Y*|H m W0Zt^ RORA|tOv(xytn?5~v4YinOLӺ~Ǎ~>1 +Ž:Zl2skNwk9Zxb{;nj~ޤZ9wuƒ1_(q~:^}a{gN]SiKY"jX'u%$[2r\5vNJ w~7=tBt䈀YH?bZifF"&-O"0n4šJMɛSjqThJ^6{Yٻ79;[tI9_?{8e2dɷwM{βR{N‡zzh9FM|@NҖƕ97_AC.;s_)eu\]1λ(s>wk$ϼr7?̦[ݟx,fRYˏ9qtt귮zǜ k/y\ifn +P@=!dWzf3|sI7̑gk:oyԞnWyTE3"j'%d7n71P+CN $26>Chsz0g0z= 껆ӎfdu]F9Pܑt|kQ mW==9_ؓq vч eHdsrɄ@9XdžiNF @Q1y".,MoWVĐw=5yV:̚ P2uvi A7@DSV TYY,到O}9ZEy[ՖIlͷK̳ѝu;luW ,Z=z {m323ţIqCh\HKn 1TF(DRB(Ɓ8TI$E>9ZP5rhyoj7[9Аj;?BWaEDOŎr04x,nInNyF)q"+E2QRF2$[No4a|;/*-t 1I *3BY6T9GBDr* r^\< E{&i{@!쏒 ʮrT +ˬgTTʐAAykdQ ;#5s"cYI4HHf1z@V2HutLMgx/F(up ៵%NS1{_[1Kkm$#][oG+_6H]}6M$8,F_%c^DI$EI-idObKg4}|U](Ν-ˆ?Ȟ+9J6y+tH /wcL_/t^t,ާW;_nLTɦ 6-]#=g);82C8pi:񣋃Afqm5KNm`=w]*f0m ڐa(Qsjw~jY"ԵdKo?-ʢׯg'52v8M>7р˧oZ(KZYtq9)?p Z)֯hq`~{q3ҏ+?~8/>xnNFK2V9q<wή9ƿA>>z{ƫض浾6$N #FׇY'Ć1',X4dowi(a'^\񗒄laQEq2= ).62+;R5_p<<cb?}>Ï?p?@oۦ ŃM@&pcК[ (\b@%oZ\Y8ABm#km6U,ql똭f"ڇarme.1, gK#)13<\C*PW&Dk$wdEu>|kյE zx+4V˗T?hTmh:(C`'tRIT} !fV[akqZ@*կ`O<"0)_[51ȐQG<'[E8vmZӹ']H*t3ni8caK4;i~pI nU=:z&td::aBjTG9Ԁ@.G$RlA(H0.sJ}z qmQlLe("`-H74dTJ V@֥ȲRh%%Yf- EĐ  MvtH"/ ہ M zjeXbߪ r &!@LN?nx:+l2m܈޺qr2XDhc΋eu`^r>fff梊G/eLQI;o50Ua CYPYRK^7ں oD{4ƛ~r*gKh۬{LѠiy`Y>ZhPy)s6G>TNTm+o,T)#W}|e٥;5JX)iPT=TC%N.iIpSSFkd:BY(c"Ι-'`-UQ]p`682&u>i0-K7`FE鄘o&~oc۫?'Ss؛wkWLlta:+xQ4׃Ry4ʅV$-nm{h=3r҄ӔLWҔVu)ga}So)'ߔӅcEK]?MG &iRJbҍL64J6>G#6=3ϼ _IDgOI\v#rz] q= ԇcuBܞ@D֕$}49тRʹ3r #gE|.~'Pus#*{!V'˹q@:LVf1BS\e69x# l2mk/]e7:L0eيf^b\\Oi e#!@jP[0 7!Ʉ7!s& );aB2T8R{OCOd0YފvI=ڈNf'lX"W\|dy*:tREOviKQC @Mgӹ\R!Z%˻ 8 t[2)aϟLcX c$%3U#m~L5c|^C0:8|#UNG+jx Jc3k7Yݐu9!4ݫ[q]/!a* wJBv_An#htֹ5E{Me#gBZ~4untO_l$nildrxќ CJVl}/%՜8ZhH$!-c, / :(M,X rT reFs%S MvijI ޻f@3{fL>[dΗyS7擧_%XqQ*̓)Sb)99a:nr9#^g,JNi V%e{¼ RYun&[n$J>^AW̛ oǾn7k6KoH}XIm$kdРLa/%jA ѹl(saAхQ3x-e_Z4"s6;~0-_ z9:+ͫN_ 3fϘǘ,C33VO;BahibBjj&u'Z6am WxzqGЅfZ z)%)wKM|ű;Gb:K'q}K=9ԁv ~ĽcF uvO[y[v'/ZS aej Zu;y0'Ǽ'Lp>Ǒ1R/lqgک ֔uD tңynųO]ށPԴ?qx<82 {8.ABn4Y&X4C=BJ!++bҐ}q}8{pB!5f j*6? m\I`fҗw5)Ѐs$xDŽ`M|yh( SDnd\H*fBfsPhM2 &}LV,-e i ڃdrڄy٭|}-gT!ɐș&,|1tךws]oLaF`#g!9h9)VjL mia f&ps5&0oL`;ט;ԘhZqѣcR1fJ`)/A8mmjz*JnKYji49 %m$&sR^Zك(Plsq,=f xzғBXt0g Qf{ k. 'sqcف %s"Q/a"Ka&c,(ufLɠ>s*Yg-_} ly@PKdGz d@J!ibJJń--wHX% +I̐xbFF#9g,C!D`Ld0t|\ڠiKD H``MP+FIOT!+bx3}'Z"W>k`at[)iLHd!  *Sq+h^q+yV߅oߎQZpȂVe;rh=&5hO$ '<gS-`#v;9Q*J2Ї$Yo&(ō/d`ҽð"{X&]wmG vf`2kYH$-ɲl=bӶ0AE٬)XyvڡEЫIXώ_-iG'>>^ $r< YP Ǜ);C3_R޹&$4F hnTZ! A* *B *B Tb *p%•X WbJ,\+p%ՕX WbJ,\+p%  Wb75%•~ Wb᪘U1+p%JQ+p%UbJ,\+p%•X WU+p%•X WbJ,\k,ԞS5슕]+VvʮX+kHeW[UvʮX+beW슕]+V?+VvʮX+bնʮX+beW슕]+`]M! }!YI͹K@"14 i4V&"Rfs${͑O`̓d)[KbJ! Jb< @ɘB`E-Gx8scc IɘT&H@xF}ybns+3i6qij},/'/GJd[>]ױWsP_6z!'6Tǀ'g&f0A2]>$G #F:*Q)Q$wAo!Z橏Pok Tew8ClivCxqroޔۼmb,j cńZHsdQǨ}D@5V qzu~)ZsD S'zP!0ƠSbTrKHn7FCm!/l YON^ 5ܻ;nN'~:d[lk VC feqV_-WxkUr=h˒plQf ^`A .%`cnWcNk <¶M$XX \Mr̊AŇv:/}5AzyEިۆ[ʼB[ {ޏWo01Ix^$+T0+=;۟G ﳭY*N74UHg3T?~7'w'|2}ooO޿?{F:=Z_%NݫTX߼jaUV:|z/5}^*05!`un$#M2jF)(v-4܀vnsB"<hͅWebN~.OJ_SDZ ZN=±LTMAlE3J  сU%_^[=ds ["U>@)51i>!ך|Ԣlp(߷w|nBل+*bc |V4:`݃kՋ[1˹J[_|?^Јl 2Mo8HVu@D㢶 @8e5I=!f!KqZ{HEJ薕ϵ\#"PI{UhDPt.1A O(*!ɖDB\kJ=!&ji.*`wItOZ#E_Y]L$َV %\'ĞL|cyaM8~pqxïCw]˻mr`ӣ4Ic/a0g QP`y#5o8XcNzTHHd D "Hv;%GuB(3D`E B" va7\BHtILpd^cM ' }^H͞ =b#_ uh N +q*; $p" BƆ0bM3UcPeK}ށCCS9@W]R֦>7՗Y ߮.<V~؛sy3BQwɭ{ݻ|88\Օeφ%rm{7{M`T ki{o[ D{kd[ !-! C'UP褲F;J ]u@Þ1NF8w gYYF <ќsyZI)BD OL3Mb"bNu8wet0:z.|BID73BDN nV<q /+\~Ncb]G3onotm˞3|L_3ULY93IMd0,<[$4@4\@RPE"'U }H@{*J̇R>Jx\*DZϼvoAɘtD %.)-/KT)!'* bT-Me80КJEy:鉶6S8J8;;<Ǘ3!6m}=t;Շ^MϺgfss{ ř+qgAqcYNO]P))+Jp!A&MQ7G u);r\=" ɡ 12J1hښ"uI{ST)ƄܢR%Ola .&v g}z<i Žկ? w~Oѥ{_WTriy *FT`2& - @M308.PEȇΚc\k\ "8tդ+`$SaZL^vvy I&}\jJBۛ0ۀ# ;xN"0oLM b*^}!j8eihnR6\/DT8WcX]bhsي]H-4Pn)<خ<UL!G &g& B[23 e2ݲ#삉J _3/F?_~~x4i5tq߁~-Y)bvmzttW̞]yaH};>u{^z}8jrҮAˬ콉F?c0)i|u>=g9h,w&'Vu4A~3/HKiu^ (pB7ҿW.͟v5/jM~j9s*7qnˋlD$5@AB98Td~3SŇ+RŇIOJ&ePEQC/=7hQlH`|%Iko;\)bf:z9p% zu[A˥˵@w߼+$ojm#/ Ȯ^L!Vtd\VR^lH ]g:^.̜>y年>.:㲰I{H6,7%-uYF&.-yOYBݫOڍŮ\.jS׫X"#$\UjeW>PI4Cz0~[ w}|OI1hlF7Ѫ@u[2="TY~ÎY 0w]\ 3^~U.ΰ *9įSo> xWu%=BH_,2ڗ/JdWt)tEf*2T\ $dL]7Ӵv՛i/Gt] mE$+gU^G,lӂp3ыG- Izi}\Rϻ\Yyj0je1R1NA|2FI˜11`T\9 5:\r)6ds^OLYɤO\FA2( &RXer!X4%17be/ '8i=.rF2A>T>so,ˬgA(9&;^ENEZdY{F5s",$ѠHYT@ٌΐ$F7ۃ[#?{mυ+:X9=f0#s`fqiKz_yFwK9\o^]I?n,>똚ZGMI>n%{RvGrJ 'qa]OFKe8sIɔ`, l5l5KNn`=K?yL`fqB\iA|z^( A8r1yQAW_uP2_%Z7Wj,zzUuVeYJ3At&&,S8zEt Imz3ݛ^Re]'~nF+r\?<<[YiW+QD6jZw{ K;H7mgjl>ŶӈYۧYWd1 G5F U,wppXtn=`Ϻ䶳nvIR7-,H}9KWB+^3ͪ06k]: ězX)?~0OfWލ..K˺C)+BYT >aWzF2<̇؄E~j_fSgB3~mopǻ߿{LS(8"(@!SOYS-N-Qsj ylr[Z,\Ba4`۾pϸCFغG%>>6g+dP=hE}{c ~יs aR2%c7[7n<I9lNR:Lk R̀܁!(-c4@&&>Ω{ݑV٦,#Osn& GιM +c"Y8rǐ(]e-2B5jWGC7FӉ#lmeQOg=>IZGQjI/]k> L6Kw*EZi]Ԭ~h C?Bn in ^TʕuR *U^+PH."bVzPh!':3kIφK|1`(\ɨD&,89Ռk`,FݺĽs2w_k'{w\RY:q?L* 8~kwvun3<ΜdBUodּlE@.Ga)V "KYYlW@\9k] sѢ-̃s\a  \fAEdE/M&FȔ#(-$VF١.E|+٬*k C',̢avؘ8CBB9?N!g]&OHWMob> yy3ƶlMJ\)ʚ s4D/eG/;zRzB]DčQ'zO^ CTsƕ^~L26J/MJ.% Y CܦDq[T"ɜ읕΂ 0lL=cfW 7W?tugp&yԛ1&דiXYH6{C.j[tBm2 _R$S]Bv]F3\8f:ѻNty=9}{5Ȋںȫ|-kGևhcŧ}WsGGü_ ]eG~qk;}7x`:)pi'P\uϭ͟7lYi#4o]]%Ϗ;(7YW"$YsJZ^()l*ԁ[cUI{. OA*[h1EP} n,pDMwGBFc޳ K#, INfʙ5qe1qG,Y18 eX(:,|T,HjzU'_miុIGo^ F_5q0d&&lfBfS!3Db`d>&D-C\Lz/()  ZE#l4=6 eKODc)~a\s.ۂڍiǡFmաv`wi]h uwNVyA&CX Z$AuJ%#1Cdl3 ac܏S_1`l "6""4!bME^jO*r[(mֆh<D72a e3\ҕ]-NzZoi'ma` ўģO=S RǯԿ%%кqC@Y4TR;/*vQdǔVOUZ>ui%d\e'΃Z2UR$ 1An4]|#QM-S9jByGZƎd .쟩`G}hA5N}jPf e׳jJU` RZ\URBmHp9k:HQQuED,sٵ:2EsJ"dDA12{N9Hƛm`6r,P=OGSMVgNt>* P;0lL=lpMI>&_ im>:E};3|,tFڂ>%o7T&6fUiAb:L|Հ4V==_͊KcFZǓY^4H޲ SKI޲KHdq<5; <لR97AyT"Ъ IfDDŽ坁e,0#M٧.jg4+C,ǔ@* "/LQygc%0ow9]#eOcgr28CPkYi@oVJ3iI0Ad2!gƅf_n-X RZ.).yCp7!x-qтE YcdxbG][o"K+p'KFx#}O([,\P:, 000}IQ*e7/oy|K߼]ҳUin~Au ()+T|Y]W\fzpȃ_%oiRB t4ҸSOyVa0i3/H}lzgj46P\ebjb6hW̶Ot`ڡގ_ÛC-k%ۜ?'64{ m'1THBXBZd0ߡE֢9*Qidʜ7@]dU,Ta*rdAdEZ[+7nQכA]WK:3eI]c!\(c&2t)BN!FPR``V !t3w}P'`6Y{ǻm{w$'i`| K]PX*|2lJ)012ZP-{p(dT2FJ 2AF B1juC>䒒BH(pZ ){›AISțIe y_865Lw.UwP>e.-L?dӋ:׬zЬm N(EHM%!hO8zIٜ k>ȁn[Qȅ(AgPe;%dҕ셵RhQ[am&g;,eF屓@t[;9ż?漢,Y*!xEԨTerϔ1 *Jy"F(: )z!!17V<ϓ^rt9o.5 ln8IP pBކq6G&ug[^G Fi# z;ԋR @*+1G6p U(6nFd+=5S}66/;.8W?$#(As!fSf! P'  ԚGfu$臛ii awl?A۠=z~hQGfz`R/9ۡ7 qzނoGo]]λSe؞On[Wf e[V՞Cӹ}ʞYk߽UYevI9ؔ3@}f2]F_'utz -˫iшY.oDwoa3[ 7̺ vuvd╍Pot~`ڠһFaҧzVHm /z!dhU^ Z_s?ǫ<\܃)ӭF'_L| ;:9Lܘ iŁEX%--NF'yNT,R]b 7h1i^nxd4&f3ֽ5<-_t_8;ųu-TZ UG_]Dg.@ʉIp8m4q`nez .ySbaEM6ц;c(S6PB 4O}MꛪP6ߥtcSKF^F6uWuAPKo+SQ80'TTd*NJeI ^TT zA9yB dʜZuwUDӻw,:n *0ɸ+Wpy*J+:ﮪ+ύz?!I*0qW⮪Bw]U)e GwE_zemWhYۅ*\ln~\dR #_fV:OC!=g4pGրvu.ק\΁ut7,W;M?O\}$y:s7,)%wrnnzG/V|/cL(Q U%oGݨ:L' ƣ{|2>Ct1fq߱e\U>[jj4毃5ZϋVQyᳪ_A0b S C ({(6* l>ZŕTjVɮG*>èJhJhg!%O,=5U9wn&e RdQ<֤ٱ c[Jw,E $Lu2C  d$CFe)QƧHlҗh(\&J)i#{m]9भub8SͦmȘ' [RmW3`u[6f߹*ȕuY ar{ IIS%VdQm&Z.Um%~> ?b2 X m-A} NgL2`ջҿx&p:%՛'rMkVڇ ~]O:f,f!9䘹@VIW p E3 u6e n?7ƽyۇ+oiC 5<Li\ ("S:&1U2&!`R)M"tra4 : %X zqofm&^9軾~{6Y{ǻm{wB' ll⺓!i]^]^Wyf ͋UԳ'OkĮ'Q4 %Ԇ9Њ hyާL$ylCc,`-et^I\AeL3%T$](EɃYgswfwzW?sx6uw>b@ yI%-F N(EHM%!^>C%7Ip k%TNyr!h2 awuc Yt%{a:dT%rߺF8O[ZfG#?f"|$@UZ/6k5@D!L4G1F̓%$BDH\D 98 d>jMR9KV-βblwNz0GtԊgQJ*T"P^iSEwdJ|y% {88V<+ĥq0iv\xrzC\οZNCe &(v&09)!G7.3FsjҶ qߤ1mBff^0c@5VYetrsrs0CR Ȳw8kjN9T#;L֔ɢ-6w3s׬VsSe;\!`v|dYwFNY:xbu u2q.LkXzիIJmD=hI- n;0,ږqYgVbs^II-;MJng%gr[m}SOeɔ1mS]Mdfbj‹0[>7P4[NZo -}ms;m`cLЬI~>=ٞ\3AV{^sKg&&sTJ=e]7]q`o6 .dz.;eb(N:Xv66;ݼRs0Uw;Bo@.;e\rI8QR͜R$QMAӾr1*'e\+T-~ezD'OP E7Q&IX:pJxT0)(J|J@bG$ EI);8>rǍ+;Q9ϥFl H# hj24 `h Q:wb)rv 1%otc{5*z-K- e xVhɕDhc Pc.'iQ-2_u!c5U Td@xH3eI!@)DDBm7cPgfPo`vuN $Q(JC#@dR# /1B49K;oZv$R%А=p)YE'Coɜ8́" 999utzj$L5־jE[G'$:QN'jrAh='7Ye>I&Ra>;?lg{՝~7+ۨ,̾iktק˫ŏ*cSb'N.pU7ƥ 8~k.!q|v wf{͈owo+ ]O.f]gQ 910 әvfEhß'Mͷv8AI# zuð,p00Q8yJ>nډo_'nN2y&׍^1TͅJHX"}4We)=t~KC;m`~M-z?_0|߯}ixvjy)L< ~s4_66DYtk S{zr.PHϟޟw?;׏|w(gop->I9:ۛ"@=7y)54͵eh(y׳ a\%oy͸_*vf24*D#x X!=2m}rE=‡1[S(%Y5fe<`hOٺ?'?k+<12`_:Ƕi4R_)}F 48k#jKrs"ZIX&ۀ2p %uOi5qnyO.PJML OASd㣴h[)19&IDM"PGU'jيz񕯖n:[]3[q:xفQ;u6g?*ˍ EC")ˣ\U9k;̃o4p_g1g3ۀ *;$z}!p;Oj#[O}A炘;.xtR?ؔ8!U*DDZ2ʀ)K:x^)^ J4D ΀*;th5 tJ4X Bؐ-1LL451'(GRs )K$~69k5)pL1Vj`9o5~Ӥvk^lϳ:MSOmM;wpKi&֠Ǧ g얮g]&MT+tk%"76MYlB۳} ?>e9kl4 -;ټu.viz|۸D:煖a<m诽=uw0D|=ꖎůz76Y횋}~̗>|hu3mv,ࡺbOL!%5V*DrI[ƷQdLmITmW naSwKHLKvUl5=9yy"OFZB3m$sTP  :rbec)5O!60cexT J`&@Bsh0O<ұٮhˣ)S;Ea-:!MONJA#%_bM4$I濫7':4zm,l2kIQ e?VPׄHbi:Ŝɒݘd e3"a1rƨo ex,Շ%% K&uT}Uw]0L>ͼ/)%M5"h3V63leM`\'wL(RţKr/VHdVo@ M_Jy(tJţa[zƛi'OKhߟw1 _2Rh-ΨoT%q`R񗤵z"%+Ŋ4IPgW$lઈ \UL:\) ,;#jUWs"S"pvʖ>kB \qٰ"-S+҉|p$3 V \q \O[U WwGv>j@.o'<*7] kG ª⁉FWZ^19OF ~|14lt  &{VRO?o}+oQ{Ȕ#N:@ 5'Y?`> &1ւl2J˖]]Yy iFrP5񃹾8ܚ7z<͇aG5( OoFy)* {dn>}4,m;,X._~|wsR%%I\F,t1S7Cv\HO5R<6j%)~Jy?i+yIL'AGϥrF\1s.pU᪕pJHpUVS"=\E`Wd _?86{~|X?/Ȯ.I5*rXrÄDJla Щ"ܗ#, ob)nD6JR@.DљlF{BĽ))ep1hknf.zQqr׉4b}VYOV*{g˪eej wX ʉcL0Wn*πw =GX/]k[  a3{F\[ ף[0Fo: ϶\ӥL;S{tYͲMƆ FmW-}9V1q7Xg 5D'hI8/l/%ktQP~r)M=~7l=,\2No4J94)gI}1i=*Ux6wYiw @EdfYܷwZڷpzf0ڐgy \ֳ@c`haIP +"J$8UXb/2LEǘDR 7P2$PM@ГmXATymeI EahhM>JSr%˷¯6\[!ܮ_, |(-kF4*IDCe)to,rYpogd"H"HǴte6CN 5Gn5> 1Fz!dNUYbc҄ lSk &ɼ8;YkW_3[?%Bli'@͈2UsR\-)9)%d֊!Kd4*̥j͎wW(lZCdAd&1d\^JZ1x.dJBpjUZa>OZiR0p 0紏\c6<`՘V!f \̄8IVӉU Lc.˴\f2\{P ?亡Il|ąC֖ň 'j4!)+yo` B%JrXl2aEX&Ln>uTӐCV7Aogea ~t?|}!ҴCV)':. H}^'GR圏hvt`.2Ϗ9Rޕ}f2b[cZ׸J\J&Kky&]a,1[@2(cP!/4rIxJn"u=%<{VB/Z01 JM_4>sWVO1i[>\7=ųua<4z>гwt'Yߡt3?ˑKԍўiT9F{EV FjMzdmYNKiO^R `R{(|~@J1yIXfl JÙ-9{*u Zc:(/:*;}:!I P.)#2detdZYM AW%10qZ@O;f25dwi3fr樴KRbiZsr"/ FE9icA$vϷӮ7*۱ov᲻vvClK|`V?HFjlGƧYʫ`a58 ee,=>{Um0ްwx;b;< l/4~4| goE\8 2pHȈ)5N3 $BfhXIcLNUElieV4 +$4 \L.rLDʈ]M툍q\s.Oh-8UeV=j h,jxLXs]pE { w1%3u36Z u1?E^t kbb$l0n*9p7fY L$ge<&va\XT/C]Dvi*Og Y܇i4%.g_~DW]FD]30oEҜP6ԍd>$aRmVhSB(ĩIz-JIh`st,-&Ύ+=?f;L/7<÷U=ԙ"kLB#iFŕWVKc=S> .@|M&ng% oj#+K9(T HS:8 YDdH],ٞBzkB^qZ3NFX. K]Ԍ+dk2`:jmCzHտٻ6dW]`GGU? Y'{lYK )R!%[=×$H#83>U}*/\C7ڟjjj,d\HJ1'0%fѳ2`;EjWD;?C nCc_9}R̅S dE?[ȟ+ %&r.{u*4NdKzӠw}y\,QIݠNʤ;TJ<];Yʂ#n(t&84a]S2$QjzoQ,9ƒw3B6iC^Qqu&`ջy0r1Ş zi}]ڔL{zӴ|y|y~q|%/_&GaY&R6N\\qV6]OP7nhty:\KsM}㛫˳0 ȁ s>y?=kzVE.h ^ki [W5|]5b}5Kv5C-aLQmK/|4\:z޿8c*Zg\WڶZk#iRELCJH矣q(E_Mz^Y;G7jVMK&y5]ǫO I([g%r$+TT0+='UH%*n6{qFJ_x{o|W~}swoO0'oh.  :@έF+0`=t?Cg;cMǷ/TX6Tm0*JM WS0|2  &-;Rt6Az9CNN  rqe.׮YqmN/|%]=+?ɱ奇ґ}äN::Rv0' )qAAէr9!XibYe:EʚqHa{lW}tJ lL2 R("3ZRV rD@TB`BeAKȅ肴.E}l2dA=Ƙ 2Ls&,tuB<|0#*6 IU6 =5 aq>PGiX7. SOal885*q%ul :5K;x^(`ko1*'z!h"x3jj]YL]^س8KSi}>Ӓe/cL1̐$2j+ALT .9#7e̶[gew>]!{bh>?L캱IVf%#l[z[|"jra1^rVY;u\t.l "g -2%{:4Xٓ *YΕ$d18ll099tT9S&g}z\V.:rg_]|S|ODgB>mi|I.# d,b s!wB\H8 …d-s%g*2Csa!$ r22){8r @mvFnL0H(YBbtGpѓ8eQmglK-6v s4CZj-.ۛX6`%v̖Aw0+7؛fqLcb:'װT jZ-Rj{OCe19JDx*<4p@ރd*r3q2zخsNZew+ 't%toM ~YtoN?+t_M'ϻY-f_-j0vv\zM/w~\ f/jvwQߦ\^ z7yX,łVGw=E>K>_/W;s ͜:Bsjz2ri|N[p1'0x8 Go,Pd:'rgӤ@5)H ѡV0En &;hq0ͤ]4;adj֭k#@-ձ'C;'nvZ` f%|.ⳑ7WLNx-$Jc+OjLh;;m[ܯt\1Ʒ'(fܱP+8J nGr påC*]:et1{Jx^LO .E38J:\&#նelM-rmD){Bٲ--ܩ-DW2z,;yrdOl@X0 2:lх L̄'Bjm$-1YhbK#3P7EBb6ҁ̂:c\moK- sy(VP -[mv`7url-q&?XsmFι,0xT:4-➜^J rB,Er.XP" A'QtƌDrl[~{X;0ڸ_}CZDl"bg;1.B)Lљ^L:DEC&K."( ycd Mk gL(*!Hn\F ";I $rɄ@-nI~xk(!:[%e:Ń6Y7zrJbgnm>Ra_eF9' SV#TYT>*OH`S8>,~ϛa4Sikw=*xNYm}+ޫhsT5sJN{N!$s,M am:ƥYtlF[qodLNnT=oQSz0/z7gg}a@n!̥8|d٥"hgu,3;Z EtDgKBn鲫,Ty0ѸQ8y%.>=isx962ն]vs3!+$7,q}<)j4;c!0w?~7 )n'o4ɜ,Mr; ~ec4\s 6wnRq g?]~xuO߿!|ɿ~w?=L듷oC['|/s .{=J!ϛ7Hi&MC^7AKvݿm6\mu2*D#t O6sA^Jcnc/;Ϧ8Ͼ2) (z4+w>?']cN#~VF$xbd ZstnOo(m;udFs.3'"ZIX.ˀ2p %;uO;jª^iD7׉L|rRjbNxN0`+>F4'$!4ήf9WYۛs=( Y[6[徕$,snk~Pɬ Ю_Iy:+VV!Re֕DV.j[9%2OK+MCW~t϶V&pzD0fUN`j %F7hsoKHed<*A80F/`m8cat'C?L Oy~>\M`Ŵ[?4=::a4=_H޷\}% MAhԌ*!4弒V%*R̲p~UNmd)T QD<,6z Z`< 12͛#%eK빱1$V$AUI !=#:j>12T'Vatn.akʗz}&Iz[q@ⴞࣇ{L;30ĔqTT3*R $V&z_<@x\/{x DHZE&zƒA8 V7ߠlҝ߳8_5iR\LͶmMo$=)2{|/9jϭL91ëY^"^I%.F򬀿@Z፪ּY*S8az17bS݈1O}#<ٍdPSw 猞dT K{>BiЧq0O`O?ߢOB~} f_}‹4ȫ'ou¡G?;7( ;1_3C$>N\&p|I+nm=NJt=p]Ԓ)QE^g]%x9P/ڀ.N6d&rn>[BqWwy >|6H-mwӯ_~U 敿lZ'mh_Tt0χߙ]OFӹ} ^`r=O5S`ٻ7{:GOBԫ0'h>߂)]Z28lw>oO^BEBTWKQ kOYS=e52 DYϼFw=YoA'cQ T8T]-/KWs<O B&Gg;DxS b(]GQhk1blx:qWhj>~Jq|5oPz>p׳yLt< +9 Y.NꨥǐN7.'O c]z=-cz{D ǫ{«ZJvCwj*<4JP+#(XN=Qԓ<,5>IGR XKCrFIdb a炨ȓ֖0dY*9O VP pV ʂ`0JRQk"tD %Y@\7>K0l^/ӁUo3-Kd0i'C&Qȍ(aJiJ8K+A;Ou>'wiq{bN!Y &p1;# $FJ)^W=a)(q W)em5LH"H _c&qN5Sa}~fs_iWoہ{4zڗtCJ">`8֨ T1bM3!?~L$kS]n> ggQmњr+'[=zېDŮYC6Q۝Pt/e!C9bY}5PxlY fX.hB fA ~zwhaXeLݗqGpu5>hKnŬSqq1){|k KVƶGZAZ*dIt|[U(?zNٯNù:-dBKKHE'rҙ*1ɜA=PW.=e#zIEϥVR(U=&3As "ɼB <kZgWi{mҭ.98yc':ϫ[5ö 6=(Yߏ<㯵oR٩|^ѡQmR@yELa)]t#o5;׮¾]<1(``#1 \pbB*mzցv\}.&S@m)*mw"ݮ2l˭GW:KCf ůʢ%k`.7CJ0Fq;eԒ˩34, cHv\n!)+oF18=Q\Liǝ#"tHNĩo"겊H\D P8*"!9SY"R9CF/G8̺z Zw)+}?E$6`:tKi,tj?U8gxbC< Iଉ8R2*wBB{!  ^E WhHr5)jeU!%TRT1u@W1hzT!zϼ F%D%s^RwH֎u9ۈns6uΡ gt`\HU҅wvMj X;%7`,_Tef1ΧI˝!u[ϝ`p d!J_n/kMi7<I|\$`&frVʨ @:c*P:lk2kljEp5wYhs7xvd8^oK ͸Ve8x"9y>+^jPzWDi5GjUdO27nUAjhVߢ#:NJӁą1G)yݏ%b]gx6dY\υ:KɮzoZI@?ܧPAǕ!P JJ'*"(/8psp tOް/^IUpڒhs*B@9%4BYgjLp6CS& )AJ%TrJGΣqQ;Ca X*qחf-&>s֡֋eP򎣿LH{@yU(82F8[ -s'2'%D1VX"]vm9CC4ƕ H!0&>NH kC};sR,$ Jp)7$SDA-9CE&!KyZ|1IW`z)ARSuD(Cy_R1rUոGy6Ie";e"\M)aIn|e[͂roȒQ[]'UҔL uFxò=MT7 02.K}91ߒBSS{ oa"cC>94t!GF 至8LG[auUQo?fpE$zeBœLq^'>5>D E!1EJ Svpe@nj.#UffSm.tӓ#]*w2O '9+=~Upf˟=e))3QN4JuxtGJ 6{`<qHh ؠBeRR1ґwe>Ɵ!"ޘɰX^ 'ՙRi kE`z7_" \K:J-e/P=1fyeqc{q?ua}Vy#:W+9bF/PH'liTPX|{Kw^vjV- <]Wm ~Ϲby:7ٿy&!k ޔSwqݗr.-˩s9+,f"\cӒ'nɛ.\Bs%dtؠ!d$uzVaEx\1ɋ+TVx z Gh˰zg񸤓Y&up\~u XqU?NYToU7_t_M߱e]^կ٧ܡoxyױVPyoLi0 b\?H,mJ΅77W83'.#YUOJ㧵w2>z.ޘ+ZR:;Whb]u)썹1W]Ǧ7W]J\Fsp39ʫvŁ/ #VnkZ}88eHcFOo(;-2CvMlbzmUjfL c aȁ2L!8j.g"R(hD) Ҝ5e $#̜M);8w~mc#t녯M~nKͽ\WY7/sBm >hlLyẁ63ѸFEzqs[;:ܮvsScXYvfZĴPMy4=;YKO:80%Zv$οuQ)ADIfɦl gX׷kKkO)H{ʩd*&)Ttk; E8`+3k}q/f]VU8ѳcz?-bׇkOޭo& ?Sgzdet>_:^b}r<{{8eP䁑ܐZ 6I0sǘFdu֒E*QRqQIP62M .Or(Kbq=_ sxvDZwb쇲=[iljLcGx;N`.I6IWY?=dSܥe?*g UίC@b ^,u9?pBu'ZhjIy<eB^N'@;!h:>Nl<3M6%h-dhJK)Y.k3:s>;OMJt2 \b1INL.Zd yEZ{f,qvY/,<=YOv2#ﳧ-M-:*C>P= DJӚjI6P 1s+SFbK!X O!RȑuAK$ATV+sϬo)Ȍ b|76{$ T ,C [@ 7=k)gJ674T &tC'XV llL&cqT blbk9 iI3V1FXx+5džU?!5&j^jmb RX䠠T`سblYs7U)ЭV߯%4^<CaV:H>Wz0qpہ\ hv~@_xO& xL1{|IJCT<)2Zlcd.R~Ē Bs沚o[\I`%c\r4g;!{4q&݆N(Nחo[nBքQZ-[&qE=9Dڪ%'pEq$l Szy eW<3:i) %^kZD&]bdctGY1Xʒ*r5B>mΊm75"Dq68 W_Vz6E՗Sl@y 0b6ˎMP#`#`hUи3Ƅ ?ma;n Q(PKL'"?lX*29 YEj'~鏎5*Hwv7 BJ#R!u_ ȅmE8ϳp4[ےItZ9mY(77{j.ܨY-n|e[͂roȒ+[蓪P9r LiW]l:maYAT7 02.91ܒBSS{ od:,\0Wdl'4dWtpiS)/>7%eE$zeBœlL&zIyAŘ-*Mu9ߊ)OӬ񒅩ϞH遹z}lXf.ߦ[~`I~QE/Yi=3l6ۼ_ݲ.0W+ .;eq$QBmE֧hsVk*P6Xe3Q: V*+JQrS(P:lk4X"rbck2x(-:X `O*%Tb  h1]j ́>V;ol 5o@rcΦ{/ﺥ(=]<3sIf||RG?@;URWHZWB\ԶrR VzL{lCnλG'b>*MNbP3ʙ?G?oiNU/4M^˄KǏ~գF~r2iO(r%\ߟ>}jAxjg9nE F8}P& G(==~Ϛ39s9ต _Ak?ȧmݷCMo |۶W"ϗY!|vhw=ޏGͷ_57}铟lً=]4;nCu<^>{vYaw5I͙Yǽpׂ/lRT?6a!)0Nm)\ $V&z_y a:mӆ;mxGga$RZF.x!>SՊroQ1%$v#)9C%Q"P5p2/ #g|G-6 zp2j*,)ռ갇%7o֫qI_[h3cvMSxGϥeZ[D$U6m}+?}* Yl@n] Flt5$]nW-r5?G͞$TE XfQ>PRm(QJG؅hDY'k͓e pg6TaA *GI,W62yҥKh˝IP;EmQEڝ )j)( ƛd)kmbHj! DRJ-T;T; w!'P<*Ě,2e43xܛ,YR?KqWlPCW:R8ZYNTLA"˛q6"$QxJ60DFSB$ %TF$A 28܁ZDTQ{4[Eg΋ˆXYW뙣.:qɶ pU^DiBSTsZTh$4v-U4…Sa1ua<@XAo3U^q;3C?]ry_6с@Q *0e%PVʕ!{A> 9j} }"x?wEq V!{kv(SN &IOI+`#@c(>F%'G)[ A3P)c$ppG- gՎs2D&ksA!_ =1|3As|}dnʣLIWx6ih.~Y,fQY"*?LdΕ=K<|]nTЖ%p`rU*xVh)aci#CJSXw" Au  s6qy7'߿Ƴ/^&quMϏ"#3ВǛ ͵dhUr`\KNyɸ=.yh D..m~#L,+u#olYF)(8tpv8F_@s vwO]}P;GY[: H@ ZsʰoǶ>~4+R?)l\{1x"@:(ըhj*;K[GfHX5[]Yv<DO.PJMLB)%Pð&[T->F4'$%N-[ޑ%}md۫O87_xukz+o=ds?,2RK%LQ^t,^`% ŘW`2J \ej :\!@W/PleR 2joL!WJA:%•aZ}L* {WZw2+k" ׂm1`Ի.;ejz)~{`ktsߎcDŽޯ7MڦWs/g((Sr!\oߜ7#͛Cy_Na^mb=b{~ .:gN̬ i4`H^Ngyz&t8'}_'\OB/b0ww CNpSP`Fws|<_),կ޾:ENi[E*R%<1g r`NEL9-xߜl÷Y&*k,DÄG@JFϬ?\ƟUD~ \>@*n \Dnv"6rC#K7pUf_*SkRI)TqGp#|L}L-]L\DR; ?"z{1}OtYRB;̸eBߥSZ9Fμ{wIUz]V3b >.(dEP  y }έ?Sjk R>>փa1S!$ (l[k#W~Ks_ᜫETJ8@;Na`L M 7]JZLcP\+s2B!<|=9SỶs>EgXB Pq*5,\rF%C>-a!?־B d fa .؂4 i)8jhD qp0T&}F`uW'/Oneo.vוx9_NeTOАw5i 5XDRYKHTŜ_sgj[q4@66V Zș,c!X,7,MɹX&U+fUyoǗK7_Ǘ\])l^vKeJrӕ^gSh׳Ո SMJℐd|Qke)!TB~v%dQ fQBeزhJ3&h\@ZƎ%"⬲jc2W)k| j@ GgoG g?(g8r()u=nxX/ʟ}^ 9$5-4vػf^_u>{;$ӧ Ov|A!{!+<eʱd9+ H!y]\ZdZ*` qHRlΗӪ r`M1ɧ,SJ٧!RiUD[vUR9j8y狳}_΍i"% ndA4t|4JJ*i QdYǙXLALO}vBHWkRm +MI`C޽ XlTd;;FrŰt|Le'Z̜y3jZ6-j|`)7Qx;DA%ڌ4)%Y15,96+r=Qu#bвJˁJn 9 ۭzQjv#Vx aMLBZh?5lӦNa7$fӇ/m c?y[ufk+\?ÞPXDbW&dgI9W0X}-4#T#||Ǒ'ͺ?-Ern0EoZS7lz֛AAS?/c15@ ]o}ȃ|>B.]}[s{vkzֺs]+17(:V`hLe\`ޣ$gOQ]ߪH9|}8BiHBX* Ϲ*{lQ}Θ|Ki849~W W|;s>ߟia;ϥ޶ەf2}|p_ϧ[U,l>s))T' &'%SpΤ]5dM])/~K];Oӳ}i;w\_VYcFRe2;A"gc+)5$jX,%T0DYg \\?zvBς<T'e.9'&(JF11AB`$ͪ)w` tY1. #?+#Ǜz$JhpFlQZIi<5iXƀQ7e gdK-vSоA)]ٝ  x{ϧ7U[0^i eUKŊ;S- UJq% *E`GgYXD;}W~!yut:f` oo$zdD Ϯ1ǻxԿh{MZZ` ?ozLOK޿UjsN#pl;c-:1-P~sSi;xߛe?L1QDe3'ks&ҚWBԨH LqCxjf5\sހxtz~c%`yRHyWScF/RZX\16+ }V跳;mj&\EM'ڎƱy +z=K?+s~si֞uyC!4dh nd'd)eސfnސggހ32oża D2\C3$`MAr#BRѳb+H8@e;J@4% gcsIVY_ Rv6$cz^_N|Z-gHxf|D}R5nCJCSA`M"ZxfRFo9.;.sg4,<;iДxΓ.rS)uw}N aKM GLҜ3XaT2|r)Kb1$vkA>JJx(um ExCγaybmofkj$4(4#bv4.w["vwlo1h24Ij dYR`d |QGp}RVc&uTĥِK$uvByI[ꤥNzA<]%C)&HFn TB{fbHm 8Ǭ 8 zwSm ?eg}sFjFu+Md֦|5P-ɒmVDtl_.e{I^%ȡɗ''G(Ȉ2)5h\$7ZM`Z-k˂!cS}.1,BAZOwC!7x6"ѷ\oHX6f +n/v&aZ_~>k6䚢-SsMra* %p }~6W$ũNob,n%^K{q'r1.u(U6h·rO?T˙{ ָ$ǺPP[kLmG>=kﳯl(kDyDIFll{A_w .c.ͤ9kHsbeEQW9y~ƚp|w?ҝR+r[ l!1&DlҺ@Sۘ6k]=.%y$.'۫٬؛pk~`:S^TZ*}' ]~~aqUtח7v+=1(maoݜiE2UQznĮ`WU,Stj/x5<S1v:# TTl-P J0GJ3hL(^0=6bxf]M(,@嚚vssqhV\{p#6_[%lt GEǡvXP{a.QA1XX[Ms-W.5 GIE/mkh5 cOUtkNvǴnl):x8l8pS_?8Qq("D\Ѯ<[\RH"'iB ,>&-Lkh0V'<5X 5:ΰf Ѣ&/XV.$OɆ⸄8l8OIfյͬz7GEɡ",cKP PG+!(8zc6$`DFMM qbBac\pXt8e?1m2̑O|c7lpb ݏ+Kw?^FO#<(džh` LrbeSEt>CX[rzgxS]8W]CC(Ps`Mi-/(h-}"}-n]#+} ,8A>lO吱߯z$ EQ#F-qf]]Sj) %KȷmFΚٚ39ߦڬV/y-7WNa}PE-cʐ-MRfNyaS3!єå0ݷX & t p!wi o*<46xF*#R"%1dNFeQiƝVc  (B' ?mv[AC(0L4x4tzMYI1 G%FL>F>ɪ)WV rYKǪKR B!chR+f)"{8UYw3)8iп.t]5+PeLW2h|Q)&D+*@RV3׿<%o>?z}ۓ?>'|¿{.#GǣZbt-u5!/bAb.| ! ,̀eqi5!ϯ̵GzLN7mdL3&0lVcpe ?Y0]gexPQC{Nj`gFv(0PQ-I"WKQ()8T`F9uip l]Ke. "s/N 8s\"*};PLEVꀯd9k?ƠS6`o2VLhgV+''U:C%aޯ>W"6Z˗WWSG~+Ly}SCD=9?_RU G._>{T>-M2d L ^2jg$xYĜ*qjMʤ"PS"!ܻKCG -WLM ½9`9P|\եۇԝUw~)[ͪWd:@Z0ʥx5Ke xc3RS~ oW*ѓorlpC9?Q$J6J!۠LnL )&]̮VQb* ؁#rIϴ#x 6 n*@XpiSoo|JataD<.ŕһ|]E|soW9|&vE×N3BZ֭#?,䰟 D c.Q9 Ss#" <=; GV'|#=Qhi:t`!.D䨔xsbRjIh3!jYaUYm,Ku3oy ņJP-ߨ<]IQN)93Z&zG{Yc wϪ"UF.~aaR2èdbfUYwꦯT| \ \%rB!, up HD!JݱJ \ mD%U•h!mo0*rb]C44"%, K ;S : 'B.V'' c+b: "j}ԃ!*UL >Ф?V~M=GO.:#o`( T90%q>8GZD 0lm$.zyګ%I?_ ]*U>cBF^uU=Xoo~0di]xfŏcEIu YB0mؘjO1yՉ-W\aY &gFrj@ >ʯt޶m~hi2&hd)6TibP tdEfs#p\HafwY8 +_y do6fyMv7ι~6= (ŝ `!/1a-a4b)!ͺXK͜ hA[Ք!I:VCb:!* s]`DʩR`YY TJ\UM"~E 㲘h~hMN6:bV㠥,fB{5^}n*tcQW*$)F ϙRXӁg` ^MP$Gk&#'8>1bdF/@f"T9#c{JkXؚdle,$>)^+e3&xnkV( WWܾ F_b#ђ6v4`hܸ4S#"",PCF*Pbf`U42`Cr'ZK(IIt yI|;t$LIv ֶȹh\mA֤cSf-6P`WSsF~T\;Yi;F# S$ׄV0r!cBԏh5> 'H 0{@72FxX:q߇ۂǦ[FD!b+˨bG$8JK87*r+Cy"pH 0JD -FHQEbKFb XKsI&9b]GF,ESqbIɦ([Eb$;&#T! {1=b*@ͱA iuH.ŧm֤cS kdُj$5 *Kчp$,S ,܂ՙc{y䚱IO {”@h!QoO3Nd_l:,ZVޔrx/υMneL)wga@ O! wzRP뙢vp+&0y} lE-6`MHJtHp7)5#_C.z( '>dq<Χ쟇ҔPRs 8{|+`rqx)ZR4BAs<65050d6q抳]3>Rty]QtV=Œ ZPx;F#LJb.0ggsv!P6i|xztf3!>=1~wO˺!˻Y,gPCi& W ydBwJ^A.uXsI*Dh4$uA|~J+f)"{8UIՎW{N%]'uR. H,IBRm )(PIŲ<40@iMڕҖ0bfoj9?9t^K}~ @shXe[Tnh‘[ 3T?_?}û_}ẋ7߿k eNW/׉ ؃EǃyiW+R髹Zlu;WM^޿-*v;%& [lw/&$16HXbt)Qtɂ2)(v4~6nc sh@ +eЖebȍ2(m;u;b11p.='"c@$APeS蜺^ZªV/ǴB2v9r&"!yژSa8;tބU2ސwPE6qr}u,g^g;d٬(N݊:Fzx+.+o5( dB%1y݆JI 9 *댩 Ѹ+!Qs:ֹ!f8byуܲj>iʵD:a=J$EK0F4$7J\Y&Xb#&  ZP B2 }J'-#"R#Ğjt ֳ#zX-Q?1G2 _ N:z"t$#BG0$[]d|>%-cIJH Wҩb҄Aox^=v ݽd)sTie* QD5.jc)T(,q.]E IJg%UI$% 5Ƅļ$<谘8!ZjB.f,˚3a]3>úwG1bJ#=p)l#ʼYE*u%8!TV@#KDH[m/LzE&&%s9ͭ ,RסX?;ruufN[ngݫ.GRr+Fp]e \D1OH_n۟KWMcw[*r7f]!lj7uo oMCs%7psmeĄm[*/#zVo iEwH\]tc~'(=7"\YcW"]`Q%.򬀿Z:1;*Wxa;!1"bcGGq A@gki!J-r*N4|$p 'FuP$HsYQdc SD̞kGhzc=;Q;DG˹??h&La8a^?{,_W޹]K?;70yz&ؿÃUDV)äsp=NP/j(Y~Kǀֳz^[gx`mA^BPE|tۇVih[tw/St[:=$fh m׫ͥ#=W_j jz3`Wޛ5. V+ot^p4>G3QFÛ~&h>1KΰU[mzu~:B(X*m YM"'vGje8+d' 'D#lJP}RIxt L0!S< 28:]-L%t1%b ;qu]p+ CGjԿi캱ۏPLU=rד܎cz bH9ͭˬz)iN`9PPccD%醒0)qY0{T&&Īh)U:"D+3&O&0Z3&*R:^1qkJqhMLf~u{|Vͦz[{0A$WgYa}nNW}b\…J< A"\\J΅< )܅GB#r!QHK\宼B##p0Hd K9DLjS1NhYjc1HDqPEȇΚg\i* H:L:!'QLF32%4%%APvjbx[.]ih [TeɂR%:IqZrPpJJx $ZPs@Mr!{ cP1ӺKEpI*1cJ^uS I#%PSmT'&h^5Gm!~5$zbjKQ mhW/uˁO(+3&8>e%V m8Ϊ T1bK~IOW4s[@̸?9]zr ͞}}Fi# 8!kQS7B,2uYͺ]_@m}] gqЛ`+7ぢU&9)`]E03$ h!'t }b3bK"{e, ^9 {ϧÓ Gq4>I.L9Ќ>ř"TF1]L~S{dxufzvC_CDa#LCLwڟ]axy;NhNs_$6B䆞-ϵ*~jBA8sHB3źfF׀㗚{ x۾I,b_QT)~{#ŏ9D+c]ej,<|_1oxԗm_XWqSvaBp}.ڞzizy%ZCfv٨2Zz)EUL']r7 eu[(1>7* ɛ=8 8NĩB'/G Ch,j+b _bccaŕ&\}@dĸq۪grQʌwvOl``QaT@:  \Y+R2IҔTyGQ^ Z 4xPr8-0!D+$39Psƀ8,AFu 4PN*W&r[B,cEa)>)pJc ыfzLD*18ް8hjosi/ݦ>eM(pݩ%[?[dB(: >]R4UM7W6nD^E_Kz,7,DM>78n}.)rH&) SIA,6 X|Á!c|qcɎarhv?AWx"ά#*o Nn$|&%]'tf̋Y0%':h*pëEiX^6!NjhVoNʔΞͳ3H>ҞNyqO?.! R f/~5" q)G**5trC^n^bj32/z*Hʺ[ZU?yY$qs֐Yeߋv;FX7;wyՇ)ueU>vr[a:GbU)%9"0vHd3m*T1~0cI5{r:XfYȇ|Y;u[,-s=ꛏoS`OnSy[kcy[e ^ͦQ.3ZOx|:Q`e?3.9'qc+>ɏya%-,5"D pWEoOx] {yos:}%bE?ز.{}YS MNZBVG.Hi2O_/N=;a;3.Vy;`E^H%EUX$Iaь$Wbכn?2KX|Wm>L07F~Ջo:Ԫ7* .7?znQ; ٚ=?Ei:I=:žmWMa|{HZu߾we mx~ QŸCI.{Эx ΁0&{'1|㭺x߮xw_i:IYBSlkY;; h]4CMftHJG|٧zjw'dǨ'k>Ǐg/v 5qmx!흽 AAU2^syx co j 5>f&Bo9~B2l\~2VjGhGb'v{9pv%к浼dمD-mѧKb+0;n  >/Wͯ~Z{iuwK 6Z:lnξ ;<6pUܓm}3聾ES_m|Bn)կ/k$$Zoax{X/6JmZcP=z56nȖM=7*v-z|hֿCoˑ eF7o޹s0fub2?qeS_ZpMI?6KЃ'5*P0}P878`:G }J2 ĝbkGߺnn|>2d֟eڨR'YsҦcj5ݳ->VIL ޷һ[g7g|Bx7`NS`r}ۀLJR p3|8 &c \us<V s[vv%m=J3m8lb_܎mЙv=r{w/w9[!4R;2Hk]VD'CDdtFU44cK1P lfZS}u-&Z%Z9.IZ(IIujUݖ-BR]&+Hftc3jD38fl|C)Jc͙>bTg`Žfl뫳ji.M!t*ڡ!͑SD2 4y ~5:Y%Kk޵L2ɪb|m=dU8[lJZ~1K"eX VjHJEu'n LemTt qju]zaH9Xv E)V XT`t[ \ y96YË &ʑS-Xl<@m!<΍AKDh#`s38Q֡hꧽh[IfűiL%eV2Ua0T#xuX%EilZTRZJ/Ehg/4TkjA(-w#mp zM )} "U3+VѬ8266C&~|,X$0Li&i, hW tafP57._s* f AXfx@ i,H@ LhI@k۸YfFeP$֔L?4 tXחL`prGPAA. U9%;\P jXGY0ȱ2QhڂgJ- ڛϐpg(QHse/ xa㡞eQP tT3|H9gK=fdqaН"p Ȍ4vwNŽ0>~9YvńTE>b888_W9TrvH'! $}3LEgƙ"6 R*F nXA#[2Zt9m@-d:dZ@Xx5]9>JFOі#r2ÝHoQ,j7V5:!.eȃZRH Vdd#p^¼GcqTE%AYDrw< omU>cP& :UQٛ;Q l+dəjfBD&ݻӭ~yqv϶My\US־,C4kcP.)|49Ћ:"J,t˅:#0´DMt ((BoL(۽X1zhrATa:0] b ,ٕ;g,w Q-5YjZZbAwzH"}P<$W}ܼ63$@fJdA`?A. j@!o,[114 1!tՈĪ!&9C!R+Y<=>B:Գpg&*d4,NuH@)MKz+Rpf 7TVP¢(hJP_mI*>1$ FKiX~ ky/.`t:]ԟWoPT$J_^..nnPhFQQ!lt7ೣIPfhT5fZڴ!ΣFjс%Bi`˳_J3B&@9HJ Ce|:Ŗ&%\UQz!u et  D{`͂z{K#V.T( #[QW \!:Pjg#.# U +Y9>CaLE#HTY [|sX:X~ *ڨdracPS$'ՀD߮xHZ Bժ&6ϙT#ɐ+)I՜d-Rtz޳41(J3>>`joRD\BmED*CJ@(\`(k<0`@ZH8EzhO8%0ƇJF=T{x 'y`8e_J z ր̯O4 ¼N;4fc)*3mHB&xC?xd ދ!)"3[>BNă n=)h`*2(߁*QvW8hw""x"`ઁtU"צD ~+}u%ϕւY]ApzKĂPU0s 0VP%{͛)vH\~(m_0a`4?" T2@@$l= $Jc$^" fEI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI. $yRR$*C`^aV=wIIA@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H\|*ό!4 $>jz$G$@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@/ " BTp($P}' I !BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BIDNpcjf:T:>-'7 @CJq4ͻ?~RuH,K .=R^tїݖ_Nuܢ\xŇhT*,'KI=L{1:DH~Bxw`<=<hj~#+ZrW#ciKoeȋg7Xg65֊1R#f0 Edvԏ3r_jH'x7j rt3胢©_-Y} z >޳H/DBUg:NpcT.HtZ5ΉU?a|}^^§ƫ&ni4/F->ȯ_ɜD;X2+yC,<Ʈ;C 9Oc5R%KOxx^}>@{k4xښXw 0'-˃W-GGd/յ;p1Ξիc+rgҌjRı%'n'ERa*yKMB[`\Wv(n}wp\Ƙ$UX**q(HkԾ"h^\>Yy쬔w!dtQQK8X p:×ྎbբ/ά gSy*| WF\ƽnbxEF19=@!BE/e~ߞ4# ny;z|qW`DN^mmP4wM3ލ)0{5q*p&Ɏ_ *yvԗt<*i"VD Dc$x(@Viٿqk;vp{5F@Gݟ)`, d&Polz>U8nWl!F]˺EJJ"\%2"T9E |L&.*nP+AMA[sMwst7D kOA':ѼQݻedU-_[q$#;Ht?,w)\&ںVii9dQkحu2`xN]uL'yR6 6! d2X'"{m4oAYW7|t,)zQKSaqj KbصmZ\kOƠ'W`ϰ>BfjUfr?!-h~fq`ߴE\{Y.5ovzǣ_ͮwCr!̺6ng[U;~ ?ߗ7JRct>mfU\#J痚*q{@6B L.RGRLL̦2떪="BWOՁm@mM5uj8wnx~g5@2pyO'KQiFUkR% ..!jm ؀O9tWbZol]$%=-&ԑ(ZyEe) gc)r)5;)=^MQ%dZ2竴f [ӌ]m!o2j '3>[—ovt5 h͟PZ{)RN |0xCu,IH׸;!Sa1$dE܆qV&W:"jfc36 rڂ6-[lB}ڭiǮV[lZm`.¹`&J8k"mFιdRrh7'zz]jf H P'J 4ؚHP,1N` ,εl[g=4vc*q_,bkE-[D-:40 >3)a x&J`ė J0UF -bklO}֊h^ཊ6GEIȹ$3ӞH=KRiV6ޝբߚ8[fk|o23 "I7T0+6g7{/QU{ȍFFc0GgoHrD[O] $y P"uja\x7NjER<-S>/Slh+r1]u.]mJwS^mNN_U'%qot9 .8kZ[20{syɮ.\)}?t93=,3z鑟o&拷ma{\ ϽrK Eh_'ŐޞRky)}畄J.V_fIo^fYU0?mrL>f7zf;^9cW^WurUWZФQكH2`XezNY?G3nRnTBCɗ7}wJNl^/K K2}M*;qܣb^E74Me[TI+@Sv۟9 owoNN2soO¿s*{?H K_Z. כ]ͥunnռ.7K[Y 1AZ6n_LN$Y1H};Ŧ@nu">[qAJ-p8sF_@ v&Ctq2$x$1+#1ˠ-vbЍKeun-o5ΥgL$ј]P h%e:&/\ycWGf}Z!ey%"3!H))/By1" g | vRzRRp;xCrqMQO.%5) Ǝj3lN%Nv݊īī. XwoLVC*F [kI}?kV}ZһJ!x, !6TR(uySȕuTBYd\啐Ȥ9eX$mk)ҝR[wЃ H>gWYSO4@ <(IEjNY%o_жAd+*K F)s^z%^{)/2V|IQڄ(YW!:[ _CӨ;E1--TYft3 ot{5Bm4S$S]oь@ /^n+89yцFl`ֻ~ݬ3+8sVqf[-/Nx-EKgׇ`پ>o[6<74]{7z&yN6fg6s2>+RR062M't/e*׃y z҉`UŌpWyV */,ɈOc:#>YF2*{-reXvevf^nMGp0|`yGIh[…sm >zܠ@*Z*l 9$N m ܶm (l' '&[IFd 1>kou L0c)d`Z,R2rTkHቚH 2eb[[gmw\Erb`>v?5Bv9ɳ+7OYZ|JW|+ߚūH9BRusNeG:U6XQ]moH+B>QKaof`ѯ.lg([r$m $C6jzDn(K2̞LU2u%!I3f&pcZsA GJԦޫ&g}z\v.Z&WVEwU~YgɄ[Sܬ*->+kz9]p!lȼ!Q7 dl\'r!e]H;Rȅp!WܕWX=&2$^ d)'"H٣ƩV츍̘`P#trs 謏I\5S9DWK{ݮr 灴WCύQowc ZBKE|[Ɗa6NX^ȕy"Mw(Zz=%HCc19ϊGҽ9 $˫c6d%m9 3Az#yB+RUAHPIdcp.\l hKuQ9m^Z"S2zT#9YKq^|: {JT^9uOdptC!QAoR3ϤBN;ʹ ix[vk2hJ4Y!,LɃRAb0,X#QDA(R*U.^jBY{ՅB)27\[+!%S)ntv.[I џSʃj 많ŌFf 0CC0m* WD8Y[E6fYtG:H+tݳuU}bаRˁt껓w/2},݋{R+.tĝhFUe~Λߢ56.ȝfhΪƺֺR7dݿV*07tt949?]ZwHWWO@Zs +Z?a5.1P-+&.ҫm$~vJi4GNltN DF=h}5 ҵ7=}~m60!*J(9#(9{:nyX݈tr^ osHݘhJ= 9ٓ!-R|(QXjVx1AmNuRuaNҠcMb272mT@tJ474JKhdd12BcG\>-{czLP::C&FcŔQJ.VIG7Z6>r-8~ Ra4kN*]UIuهI[Tff8w6=JdudxJ([D|)dE"hi&g\/h2:d/-4:!kY)CdD*H- !) ɘeL{2 86B:ĹG~-]3OP>咐|8CnRWntjv?˦D\/&+@Rp$(gNgNRETl)Ȩ1|1-|D0U3go +D(`zHy5//tujH~H$l,Fy18:^a#o7هqX6nE$/٘4rѫ*‰LԲfluSCs["iGe~qؕ'PEsN3.;II"}*Q^5z[g-sJ43ltx99Y-dr]0ջiwOؾX{/ѯeϣ7>ϩ*Y.Y\W]lM=^ W N﹑suږݍJWwFc 5o ou2n #fOep֧{lt1v)j\&'ti Qn}ݩ.b6OY5|bll]g Z=q3j識İ\ loQ\ө|˃֮p;u)6"DaЕͭ<9>܇m~*7NmR!&cSZ1k]RBey* ˞EqSi۩+0&}YeL8*FOe:8{ۛ]ХjcckTgxH?hqqamgpJיC e[~QRXt4V eA7e/Jd H 9hv 27ܫ Zy7"p猉6za! d.U:(ChGr^$D>o,'i>^:K֮`6z5qņRGI&oԀDUsM/wߨLT4kr(mF9֓c WN9 2AsinUig2OO$GDPkəs3En:.r`q d 'h 'f\iYrMڂP!:&Q'ΊzVM{YW_5[2PKjOG Ьmn aM;VpɁ(I,'QK6 *PS.{41;C}; ^D#D  h!3i:!hv1.IB P Og{+ĬgNrZtJ}!fȸX@pT1ګ6&ڻ⼭w8&gV́j,3 ^Hn^ i9LxU̲N[!.)$}QjCPӕoxcJrwLND $@ޛL+ oa9,&az\&>E;CTA/eb n:=Ԥi\E?~Hu@l2&JyNo)el+بoX d.&*}(wFfl>xi*td Ȋ$+1&qK?M rT:@TLKt[ehs ц Pɉ ŀk-Yz Xc:(=*=Oci*\@f,F%u"!m,Fϲ)%#1A)܌hC=$gUI"(=J;j[kul;Sj?|K'g&G'L 朜h`:/ *A` u[ꚢv[0zӮMl|ׯB[|;7/ڿkeqr4%-*֭\*%4y':G"ӫ(+}/>?N`#\4?\*],ĵQ"O+zqRwJ)E+A| d [=k-q\w3o[Go~St]rec'nѹ]5v]-\ %^Y wc{v ow:->;ZqS5re|OՔvpGLfoDmo^Uh}_,D-Ũ z%X#2WE`Î\q-"iHaTsz̕/o ]\*ʣ1WEZ{tU4|0W\a*=sUXU֘+R + W|z.9oY|~6?"+޽qgv?~{oToΦn-mJI_?-޼1.'Ӗ>1;ZY߮X=}ŷWK9zFY*N=OklC]8MKbFVv{ň1KW5{fIubfQ:p}F_bZҘ'K#Al+tL|1eӢVY V5H2*vYyLȊ}!+BV\ĨD`s5ZL'UtDQOF᪠uPPz Dƅ@u)BF:kkd*Lv<Ƙr r]jܽ} i|]]y^/?zl7wOþk[Q#U>7ؼ=+UBh8|#`&ʊGWUUš:@F2zYtL@3Az#yB+EFJ"dq.@I(d\)lb!21%cYH<$YeHWM!]^,Tqx2,G~ rDN`0q񖷩L؇AIJ C]\E pԚw͍Hh*.Su_R\|b^Rd}CRKJHQCҎE bAwiLR4}z뭷kDGAIN-|z_9 H))r2RFc <UOGwTD>Gw3䋯I(ЬL6aVkYecɆ@9`%xuJE`WJKdN`bew*i`j}1u,Uℶ`Y>hD(Mؼ"jlmT\*B*K nMs9޷ smC[$à85)PdS1o&[lu'7Σ$ѧ|M[@5Cul)Ml|aÚynMblf &~@A ??y0MMޏ34G aEgV]m=9v5"Thwq>|ԚZ ~u?Ӯiww"^h(7!~%PYagQ^,^K7%jtEd.:@]XɠKpM"Qel{P}z󀫛)3{Il cHXaWi Nu@]T 0JV7XĽ{¬'湞(Z/?O[Yo6?xmnz8%;z>~dM]}Mzr2n>]F0^a4:X % DVx O9؛'N-;v`r~2dOi%2"cŹlfѠC(&PFLѢBʲHE ){ƛ䔖55GUl& }hΆg4\M'9\9ãjx>o?v 7/|;eL\Y#}ɬ1~ HB*)D>b:9Z%|(ՑSmQK`Lv9>+mv (mhg*Qt]ΰq;d2/n{ G"u4+zϲKRY3!!vYp6~9o&7Y.RF"YXd# })ޅ< O܅T'Br!I¢;I2&Uƫ$(gV&t @]%6Zd`d7!;kA*LޅbM8"'JK|JS9W+`6n㸠j@vPc;q `IALWëxxZr'?Mn'ãg2ExY WRm<}^]oK$ __.2F?70?_⻟! pҶ#o]/uJSP j4 hcdc F@_Nȝ+=!V"۳=hIz]^J[wҪ(](&)P2Ph,֐cj8Je NT>`Y; {@{gTzJi讓4t_G}-'5vRonOc>mCdڥd)Zlb-.vAB0xB4V䨊^sU^΋) <C"yϪ*GJ!!F$DZбS=xERIFj ҐEK'm&rBe%؜ީg^Ԭco R ( hyW`F!]۫JP5+CQ 4:r] ?gPa_Lڤ)D Vl!.WwdʊE;~:mC8~R' pmÌQѣ3z3q_z+0.VYf>\"BSdV"5P7ѝ#ڲ,q پ]&e;)Yz'g&o7e+=6_bjͥqM'qEURWTY?ڹp[/[?ZXVVZRf\8"{|x:'BוιBLF.M4d3JT7Jo5{mI323,|jڎ?!s=YGI'`הpk%ťo1-YjO0vvM/zwqI]Ȏ}[h;s@-L:8kbFgo ISv5=3>( =+c3;G$(}>"X'K .kH)@ *lUrJ{!2%DN~%ebJr RMն>KMDJůA/}@STxS?l8~MVuB 㛇MT#j5_YK~?!TDŶJ-{_WeQaVp(](x=\lJ&KNhz@eTfcɁdCXA)< *c4 #|OöƳm/V9Fd^_.ady2;ŢIǞHDN&eؕfQ(1Ŭ1$1DR>X%J$T:Y=zAZ'iBo-qn-11`\¬[<'/m9#S3oD3 #fYQi::|;_ngV36 wbL'u8˯|~'bU_}8t~ݦEˈ8g*pb5ۦi%<}I,5 g M?-/ߓydƾ^᪽BW֞4HEOO?[b,nzF>yz~tY]Y+v&<^>vXE-ٍ֪*e3m3qЅɬ塵i'xJy2ylβ@q|9mq_|m`ڭ [#+mGKh[;BzF۟ƍ:N#pT/C)-0f^/m/ mq}eZ~|h\wӪ~3cCXn-]ֻ{fَ>7o+olx!!ܠe}K_3nSeuc-.k:_^^ɲ1Ś\@qz"kJJBɓ/eer)]O(6N; F'ѳqc˞xlXi*RDѱ* clKG z׉ѫEsRO:8B!.^f=p|Gý-NqX]zI"=9w)% c& XJi@HAfcF  ʅnKAZF>gR$o>Qo Ja6)Ktě|H$'y<TW3AVO)}'[e7'U't&Gm,J'2OM^h}N\2T:X&$|5Y81亃f?dyiLD50^#ud:Y ֈY1ên$T}l"tz ʺ#* :8@'\$0!Z18SIdJT%T2#:UVFq@4" кWnB]!DJ=mEI?wEm:AEzWD.|05Kf%,60bf6e.wӧE-[CSB:N!|}݁z<%$cI+H}Eb&yAl}Z|Hb+߷z^-'@,3쮾 6:}+S3k{u{: ^+]6ڟJ`̭! H eL+R3q-tH*m/ͣ0V~X6_bWw J^hjʫlz=:[ZR4B)ǩ%-QCf o`-YR:nXjQvQ̌ge~z"9+`̈́6x{Y,]0D|34Y^Ch]M)E迯uՐ,jX> FAWmGOuMSzmmmZ׶Uk=ItR9pR7@X$>Ű>[DYTR[͇em{JTē)~$+V&JQ 5E܇I*74ULnbP?.\t 䳟{{W7_~׽^az/~{+ ?qVYN׉"t yU j*K2gߡ^Is6yMX 2fWAG29s:o6M|{.UΖT9(PUh%P޺#hAF=iR;kc3n8} 3O[ЁzP`Rn4"@)MdFwSeFz]=uE²ZD5u^bG,]cCTVJn@6[BQQBQC,PGmNkGWV}jvj[==,s]~PkWrv38`Q XjT ʔV0bҥzR=m:%ـ0Nz-q J#+.oPT1-7ebqvHIk&wc^pRzrZ+ I%!z8HVvHSBJqTQܦ]IGEdQ0A[R.k%O,Zfg;;.npE1ѿ pR&gǦ^Z]2|R?㫏0ky|t(&lvRg}(LI̟ru Ըߋ>cz䵙)z8=H`ă SȾq8L\ajҶR!2WE%@$K68jU{83~WŨNk^weCV|vSRLzSNq#S ,3r#c6q#c> ͌}fBaყ&No3^~vw~BiGFK\ 8v4`hܸ)͑t L VA!,BA3 j-$&,%^Fґ0"wDI:d.q#y4's͎}QeFm֡v`7SsF~T\;Yi;F# S$ׄxyHq @v4H$P## gf<&v<1 b/"̈;Dq. fʌj?<|x7YT>:'r{s@wU 珚_+\ݩ;\%=WPOJѹ`gpRO/ACp^_ָm1[dMfѮ}So^@}#BXYLISFD]@H0<)c"k:oP{EcĻgKY{ll!5j 1ˁD*L&wH Қ)1 IAh)K9 Q Ƒ| <ٞȳ}?ܱ_#\J 2]NW ]Ui*)|UQax;0kP=-~u'T?GWg]iUx!Bqʥ>'MBk]Q\YWנTϤ^,(#| )O SNsqEp Fb K;{w_ݑȹ0z }t5՞M|YLG#ֿN;|U=x󋱙\qc:dk9h KVin[l|u ]:g]7~zXjsv9bCjTfHr ,XT`#)E~2R _SN&+XyKҞmʼ] vTsGByo(G RZA6I ~awМR} RY'E2Bk&d_p@9ʄ4) @u:Y`W~?Kĝ%R2KJ_+LwG윜^a~TtPԫ[?| ea痠꒿?<cHzN%N<0(WNJ'JLX5KbU:`UhOMr3sC 'ߗ6ڳ :=N 洐RhԸK9ZI;j9P@J`άVjE ju 4v`dԌ"aq !PCga^EUA1`Y>Fyo=%՚s4qUc&3rnKs3ɟ.?ne#FKPaf/cFDETXii#FN7#S1R`1 D6ʌ ;6`#*$;A ,P(پY:N9lgv}u6]_+:4K`le,+(}۟LǷg8SW*${JF#)dR MJD(tTR!ّTHvD*$; )I68*-&" 脡`{!䙗7J4D{Rf`q؝܁ 5K(F[ ނ& 3Bs4Kc{F^ZbО"7+>IO'ihlK7eS%'F=[$ "I2\n$iA#EE)"q]^'n゚qbD*@Er2FSK'.c)kP1?$ڰ z-gi[NLH$h92 I2G26X+e>X)6r^qfPzHX,5 ј>0iG8o~ON $|ˬRFyg2sSҀP7d eSqB~j{ KF1{e@s0H2E3t63n*s8}˓Ԇqr:J8MMXw"$l jAzҿisxnN\nnde/:M+Yżm ][h!i=Ҳ% l2%֗K])tP*G2U@]:y\iBbk"Xu}T*c=yUgHOky$+I;(uL^[cU,&2bZeM7s`eN=֪rG6@ټ&}(eoL}⏵rp_]֡-V=#m CUѻ6I7{DBٻyYf}7䃉0CJ cJ*f\Yg'ЛBH&StMEin'K E)nFw.Od`zgpd^:# ,Nqϒ0ډRdAhz rd>ŐD"he"\I C`%tA9Q{B`#sB̆t8"oj)@m*ZMWuEt^ ~`"+ C@0{uAzi=,*ILaiO3ߤ^lT=4OBND$B SK&#m7DoX%P~ax66)pQKGqȚK=rG2HddJv̑~8ݓ,ՉI͘N#F==VX/B 2{[Q80;ږ0l'"+Jpjttx\߷i79ݘsW\'Zu.n8>y2=08L*lgw|s/x`y噶5/瓶_~8}D,eRANGp'|xəh^OaS6-<㇥zڨYq̸0Li ݇U"x\84}׽ؗ7,i֠pWQbe7<#<y<9\@v؟g6sZxlWTdztv8N>_.y=9y7>ѫu2*[}Zg;D=;:|w|2?4ϖe.gߧs2{/O^&?ɨx/I]pt-?p?)@_FҿIi5/XkdҒu~m"wN`:u}n7*=Ijk=>^e_@p{D[k9Rrr?l<}^d @NuovqFH w|}yҾZ &wIί*hC?]t\l6`e_}ژ. 5/xx1q~n(hӜ=(m|(6BZog[Fd2LR0/sDm@]J3jn']E}l;g7hܽμSwmQc_?7ܠrKC4o^xnSϴw(V[Y84dUjgVi9wΣsdA}F8_U|6*8}WK)%cZeom@"\BwHP6#|^W80f\7.l*[o*wէ=>i3ikU+½UO%C︚8y:(Nt"R+4٦5@ֹa*Xc,g,>V+UU=r}끆FsiR%LfYP9)(H.a1 2"gnl4@hGqxmJ. ǘ T*(oly!y+"g;-:t*e h}f|^ZI{T3أ [$ʢx*'cVF9m3 \J`#82 do60t/$w=T^<B,ytLKgDfdA d$\(ƸXY )%sbdl2pZǤQ"H 2 Z0&drV[Y+W|býQKǞA3 ZU8'LR'E`Z0d)92ABU9EG=SԷӏ]PN@ZĒedɇrQ{)iupC蹐(-JJ萳jUZq>NsK*Ng WB됍d& VcZY6 3{QEq^5hoU%!v|e.`US\{pARs%-iX *yK[tM%Uc]Ve-CE%5J TN:Jh"*7pA6 d~ðxX''g"8dW&,яp 5i&+/WOs $!w6y$gYΎr'|(oTT|z2S\,P4J8UI]zNz_,[dHK-Wa }YR|Ox&һa)!#A*݃A rx P`^6A ΄C@=ܑwyѽ<|=G Z$*R9HDxJ]YUH'NK,E;j[.kV<|sy)5o.CNJT 1iH%D:,IRڶ{%+vVTO^vM?7aɇ v{ҹ#7Z)+˯| hY+9?aLNJg|I,M|#]<)Jdzv?yp}QN^j?}Vvy?*%?q^w$[#V9+S=F:w5LA4 o&da@"0M*KDEW~xhhEŅɽ,0D1b5_"Gy7QqQן"zDCLAqR!8Nv8i˜;Trs7}JvxV`I|\q%n(RrПNA(yR8,.w֩Ĕ}Hb L$ :5/%"*X!P@kʂZ[8^e>̴8^_V㓧k< مקē8I7+n{j͖aIƼZyF}TBŌ{4YW% גPLܩ"Ku2G ؜R@ʒ R=!)1c 7sϤjkjl׌J5]X3vՅ..|ӂ7,n$۝gg86EƏFwlkl"FEIp%D BbhĀ$![S̊>C1JMil$`T%ɶc(H*Z]c8{w쪵Uez#7|șddAK!`Y;fw d1YcI u?EVthI$d A'QtƬ$HtVևٮ[~7QD]5u{xc1Cb&B%C,PZ<mLf pie p nP+]6dE@PHF#kmF_vc1n{[ܶh_0Flɕ$wxSbYrrRԉu(rK=iJH*&"v֝5_ud8ک]Ic"vlmo{2OtIi'HX$o*A*A&ΤulI@iv\n_piEF\`mD[9}%Ѭ9Շ!5>lG*&i &Rg);2h@B*)ڏwp[ FcHxrPiBfě TX拟q(:].B0hbP{ ^ (8Meap!?CeY +<k4"y?dq/ky{R\kTOuw*gH¥+Jީlo>pyRߖ\?Q]U}d]5¶ m@D*<6x>hg۠t3s5zNg$ȱ l;d+E*E0 #sTdX-B"ۘq bP"'*/N\pQkC62yʒ]t֝5˅R!js54`~V.o) HJ2KY@em]^hբR2ipd?Xt *xq>hN -)DKt !aAW{VC)5+Ym ؇ vs@Ґ4sWʺ*zM&!!ʈ^/|(L61 d1 #drziihkK%61A% Mڡ+XsI[EzTaWFl2fU4-(.Ͽ"̞2-xq|- o ź'rHǖ me Cz_ Uukuf2ϫ.'Sa m=^J;Yw& cwnO %kLZ&:'{s6{tٳ U2$!(\@GwEw%Oq:hwUhx^/m1;?B<!}l,d6KR61Yxf__Ok|sq~4upv3؁[t4|:ãj56XІ?^op1'4|B,kҖffIf;VLQkx,֕|4e.J&0R6,>Sb֫ ܓn(E-?N {Vo/>}3?zs9>lV%a_.EUy_ƧS ͪ϶^M9 ղs: |}[_7߾;ۃw}kS:uԓ˓:`:p~M[U޼i`Q&|vQw9%vҪ+1K5!`}oid黙C~D3 g1[SͨVc8?y7ݝec.3~v#Y$9mz09a4:K:y֔fԘ6Q0d22h3EcISܻ#sKش./:8 0Eb*1K)=!䵈Fw-.2PBT*d IE6֠k =Jo~ƀucϻ Ȱ4>8Dr 8C߶-!LfQfEN,AEI θJ%eM-[MrwI DZ5}*?Gn~'eFQ<_Ҩ]tqіБRҨQNRN*~o?. P([LZ7&آPz\vQ3˽K0 -*4l&H1,\J6JJSJ2փ6!i(bL/ mbLR"(s.#Х̢uataw)Ŋ|<KKFJhsWc^GJtC=g?Pqt:9AƭQ\|F$.#\yN4?z|+[y9>;*2i 8&@QM Q5h{)1"bsGg dtQ۠SMaА [-0z$iEN g\BLV$gyܶBHCZ9f’y^'ZM<=ARl@x\#ٜxk OyWA/}*&R0ջ !\&DDSN(2^SUIsA(MIHjNJ>dkLYF&(X2BJv;;øsjt6:9N;J'nr夜QL@JjE,$Y KI¤ PCw!H))kUHe' N&t"*-+-q:"f9`$] RE9c%N"R%sXB(^QdVMuGUO:qP'ˇX c6ƀTlm҄>t% z!ӔO$]ܻm- p8)>Nߔ)%ΛO٫.v"ul x۰];O7^" :Y=ŷճd+~\OVNN~nLR}\}/#]Gh88ru+Bxo0m wo/^s~88\+\|e"ODllo=BN=Bt ߦ큀k bmƐ+dtBsղ]N?jᔍ3$vvL!p&Ȇ$M47EzEW7Cz X+{I;Jڤ Ze\1Ig:]4Ptv+ޡ.Z`פ4W,f}x+n6mס+|rӐXSg#Wo}>'y89vqUDa6@>5NhbɨGAZgvy'",!b| =EG [c] v)Թv"Tt1S`^U; NtY$\EÐ:ΚGRs6}Bhy?u>Vc~.{ $J G2Xή<ʨ}(K` 1&lcPI:b6^J^لxK^J}4Gޤ/d,.)W(HDL:R$BA]ua g}z\ʲȊų_=*޷})r,Y2SsϒW\ѽ,Q[/kU)֢"m^,sr%yߓ5R]䳊!Fk7tR!8}"$ FL!HJchiM>^*l Dkf-S<  ݥ@\I ./DžD }]/ﯨ./ :* &gcN2|%-[V,Jxknw̲~y~gu!MaG)D42J Y !$N,4*%Kdj&ڦ#Bgm$-YgYSZ PVPG`(1W)YP8ekV$Xi #HW.UɿO^No rYYd rHyu)HI#ig@c1i m_6YdȾƃc;kc`?l OO!(W=EI$EQCǰ-q]]U]U]lI1Qr׬7R'KjkscnHxk6j3)]Ms Zl>'X(MAp/,Z )ijWpw]hev xcZL#A -%)K *,I9aA^e&(CKmYD4ϩ[~(J|a3&lA)Tʦ9w{ .0EhoqZK4֊¢h B[=h Cv0uo:`}PhMuC?= " Oj䑜M ?J<TIRT*$hWIP V+提ʨx5S:WͦAlN{ɱ6$Q3-9HwMm_#ĥ0q$ۧUDڼEm,MTt-"'%*Zs#i#D+[4Bֈ}^-.+4)-<ǟG`Η7xa99?\6V3.3BΜ1 15Iq9$9 Gy⊊jn⭤\Mqƕj-U-3hHUϹ"^tVȷ~H5N ؞-Yk/hv}Y'fǮyHpVg3ܦ8W6ObEdEJpre;P~@T(V*%@^{-H9o,(1-X;kq6}Awg[JOSwkr'-6wS@oj1 >YƃDtfL[̣1 υ^$lp -i񌧳3߲<5~B84EfE@"3]\A] xw?Yhd&0)Fii깩IcB'4ԟ/˫YXu`rfp7m:؆ Ӽ{+*)"~u3z׍q{8ol{pX#RiW wX7 g3X7%KKp^HN|EGh:ZW%ֹѼNnga]. +JƆhE\A.d$"ѢRQz'n0gbVt׷>(,SgEYP&wQm=_}t<`\򣡙'M9??ˁ}e927k6%Vڥ{MfQ%Qw?_?TTp2xDyEL{9Hn@'_R7% '\82b1%X>!?ϧ%p#P}-Z?LRBQ*_A1ӫdQ @:~,,y0 l[o{:o@U7zq| t.1,Kn#`ݨ?>ц N1N bS?-ϸH"3R0t>]>/EƘ#t9lVN_@0رO!WefZiʝ]o7dL/Q&-v&{)M,o2.wh6ojʹF5@U=F*NfV"-AV=&hd Նj;NF%3s z{orf|lS[uHK4 @AiaR&*B4hkNPe,i[#׸򿞿jCN 8}kB7sc\XJ&#KHk  0d<X2ez0.fU1)=^dE;%hP.RMW&]YYyUD$3z2H=0A#ABf-~-acy YU#dg5oߪ~QAi_McA1+RWc\Nd(yj#׌W]-_zx wo:Тh=D%z9{Q}zLaeB[k !;0S]r+`Zu#crNcoIcgU"TfK (W HlB Oc)%c(,<~  ^q/8[ d/n]7)%xp"sMT ԁdB# VwRP<C;3`nzWTP& -,),6W2 X;),ӓGa4+Xwc(|)lupљr`eן ^ttԛ 'G֒bJD]M#p9(תYLl~YqGH\R0seFUN<caXySx1U^. g7 0CGӳre\%C xyfFsк[;Gf ˯*!ž,#Bc٦)vT6ZOrݨk窵Ĥ<]fBH~@Sb'Y{;f!$b~-M-:?x1K7<`aOKNB?E Gtʄ boMPQa0rQ3 Wo^_/oN^>>Dp e\ѼF {=1?^554WMfh䬫b\Iq>.,DP.:hmͦݙ+JT=R$(hK# bBDf kh#J,3*JnqKVϟduP1x Ke.U̒#mvP* VDT~;P$b i 'v M,v_Y*k6״Bj}LJ)e>L|'fv秳؀SbRcg6(Y8SoHǓ{Jl_ ]C{Oal̓}0t8tJG$q%5i\:WlH[ٹ5!ΘI|ƸG”f܈HB؊@ ݩF.y?ݝVlO;ҼtB]"YQ)=H9 ŤԒ=g BxB3ҨtHSukJqTQܦSIGEdQ0A[R.k%#,Mi ۥC}Hɻ%$ ̂HB9UaaP)rsYP ?-k;l|r < YQymǢ6YTH\A3Ov@`uAD*A TU^ƶ#)7󚕋:u Kh^? oɢja6a? =x e'aΐݻ[$`T L'r>N 0Tl˥gG+Jurjnf䏥16Ta`*պ..6l~hNǠ|0*R7+(zԸ8B5OFS~MBqXtԟ ͝q>g)0², Mpmχ)|05T85PǏ?WKE$+В)E 0ȲeEA8c`)"Մ-,'(2G`DXgWD0UŇ٢QF%3=@UdcIBW@0%`*KɡU}D=\}pE$\%>J䊃D{JT*WW`.-MKOOk '_~~oOjj34n:eDdB2a?_$-`l`Kb,^N8EV]gi6ʏfaJ$^*b5PR!,"(lTJωZ//ƗJHβc:^(;T!{W=7O j)'7V8E!*Jj('2#33RYl]&1O r;)2""FQ4`pHufc>9[Y H.7}*cF>FC`?NMc0Ўa' |j{(WR/ދaJ:TKYJOCKM<rhzJD5Q oz$k$ZP;yq~̳-l-ϙQ 5!qqsͺqp1dF]+.o&Rr]q厎TʹVmSG\뺣(]~02샊ɶ~Xs?a)\ˮ{"F03Ņqg0/r('e'R2UqXTzAؖ>[M7-o_jL/fl%ȁ>Nχcm̮x5孭EL$(k}~7t1?5?Ttqo4ϛi-)_ 0{oχӷ]]ҁ\ GQhr'}h~}Wij45H==xv9懶)$tNaqӄ!%gZU+KӔfb#3?{bjg[_CQm;47yӍal-uwu'&.p옠fPs ocy&#sA#~`Ńy'fND v}o>c^;^ΟU%i"AUɃ)cMT)XD㝠X&4ӆ,q{_^j g_T޾vEz|P)H(ӂ b:mrb ЇA/ %t񴘇Ei?9 2کuVǻʨ=VM̼uVU5@J=K4C _nކ>bX ?OO2*csr0oHmZ <Y)9ӝ2V&9cЧ-cR>-Srߍ:줄R'8[eP>eDIa'Jj*0DQGQkoAs^Q JÈfUB 2V@i6(҄HƠqf2张ؙJh6r6+ C^NZ-0D4&x֭05[Î4wt.mIMqV T!™^IU8n`PkFr. F{'{!rvE>Iy#q%(c@c`Ov00{Рh$VLZ"$"MTR#*#+TT!H% M3Y6rR>1k|k5X@Ds m$l)LkJ, CD@h #z)9z@N?3{<\z%;D@xu3km#G_ض(a.3sN0b >m]d#x_Sȴ-'[jvUbrR`闈q!4IiMuP AǓ:4^%L~߁ͬU.r峖L՘cB"ކ\,:C88Zګ;-5{ތ2 FX5ʡR'ԞTNx,x̄WeMꬭތv>NM%=㫦"l|-s[%5J\NT*H%#fRh.o &MV4%kr_jX:[=5<蔦=e(ҕ(-(a'.䙂M.k} ʆr#` b*r: 㐻R0h)w`$K(90*Î4&0X")xB7Ħ2@)-JJ[U,T.HRdY)O \g@C,ж8-y憇ާ$@.kNR-/8zV.. 3AL|@g9mf͕5+Ӏd)ƚ s4/e/;xd`KxkuW> sVC#&'SϤR* d۔.Ib&s/etVy]S*jlYgk替`9/FUxoCÍ/sӲ6MOmǷ #mʌ5 .=l^s'ܘ䖐kEf.5g{K/C.o}Erv3;b풩;9S6ˮ;W:Fŕ^ƣњ-nt3-t~4ۭ!\U}tÅK$]YXn~z{{pfw(:s3qs,#rsg lp~NXҥLTF@,c3F:˄(e߻ւeI yɝDBS2iA ކēUڨ`Dmy:) OLb: y[bxT0AIT}K\E/@f6Q[c!CG&` f, ېTFI+^ɉCOFbT.r0 8[D%50>&Vӓai/|=ʛd"1-fܱ@+I( nGY8r锊Q]Fc3TYrA' L$Lt)bLk8W29ڞ8=c=RVӌ]} E =kr>vz^핳[N@O.'a:  Fd{lg V# f|I%׎ )4C OHPAIl2eBJF齐MM !a,&:W`0Zl~4(tzm쵡]; z'$"ƚAe6qeW[g%uu7f '+TXtɐ- :77fY L$geXM%1VZGT=_XM;vX\X U9 )Qm'UZRƀFk~zr6i1"uP.q!w { { C(*nxDL 罱Q9I2d@-BZE)3fqh d-HJba@eHYr1LJ:?He VF2r]hƣϗ{(1>i>׿h7,zk3/',!鬓$jÐyu˜|"%VHPBYd ]IQW`PiMb犠yK]"SVV ,JZC(<ƧRCH*VYͺ䲯C%J.ӓt?$OMviqQ1{C`zfJ!b;"x:ީr:mi R  hD*&eN#tglǔY:CkH;Kqܖn܀HZ3'@˒L " RY`Y`Jm7K?͹0u;xϟq6"H|L794.z8/[ςy>9LΈZ2 3sH;c+Owģ>$[t̀MhpGez 8 Xgh) n2I~py0<=H̏-~fVsKޥ A 57^RQ;O<<>*z">w6r1Ş E;Ww)]\JD=fj~~}8==;VcYA$jpqnL*KqƊ3d|tZwgfiFzN:;=d0-9u<zVګ6QD687\v cZҰ%PtkZ~0QԃH>.=jڟ9+[l}\{^iҤ,uϏq(>ʽ2#~xo4[ {Q3ˋ0{?nOBY\DOse*ʓW藡(0#=e3='3}So|>[tmuUVGػ$6l<_&RޒC.cȂ~Q:QxZh4N%X0JFr't-eZNᢻ4}t|6\"0&cDGAD$' K"BƵb, ǹwN"aG&' ڨF:Z!R< kvH{}kJ]- S=뫏+͓>嶜q8նnwԮ-ip9rQlZgҸ x IHrnl1xEC^v!>$O A"U2pBzv2rʨZ2UYc R$ Q0&h!?_7oR٥rAZ"Jm3d `{n1dὡoA\{{q5q6C^h юVo&g[f]&'/.83<LG>ńsC ri|Vؘ+c3ka!bnJcC@A@@ٚM/}^{i]CSC[v3빩/VΥo|}$Vh({'i؛ye!v!׻i go/Xzgbbe5Sdݗc".\zt`q>}0W CX׵s 7z&ő8ޣ!-2h(k^]5Oڠ9q+.xM␛& URB of1;ՠW[z X%bIb R&\ĘxFl:g ˈ`˂̱Zu] `cd,˥CjAY!f98/nmס#| mSNRI1o]: OǓ 4 n@A9aI=Njv.;5fߩ)5N_v $.?MAd5p>hl08YQ}Rs{ ZI~61m79kv_݌ qӓ9[%?8T'wmm%IUFƥnl>٭CMN\.\W%R%4fHbDY*[f40TB|=eHX z[$1'[HCI:JW*Ӧ#ISg: P\#]0Xh#O94 ClH6ŪdG"='1 ?NN< b1A"2#ӹUQY"/ _қ(K5c'it\LR302tA`Jd߁ϾĹ\T!<ŞRjg^;3*2*xPK[/E7!(KU1*ѣDPG?/]ASiTHe'NH7Rt£*-+-kUfWx7+V^7H!J*>ҭARXy8k\}q*'_LbePn!YʛԱT AqH-"MؼR%3T҄e9$tlt4w#3mY#w,:4)H NvOjnS[wR:醪ϰjuR;zBm<sUozwsUCh@ӗUը$?tz2zw?jb?eѦEyw ::VQGٮOR~^ ;gy͹'y-S#^&Ҙjt篣ه G/IotiGF07lξ)!f!wD%aˋ/[zHK#C hPi$)hK^>': Dto _IڮܒZYYh Adup8@a1uǔ@p.x. ,!Acvƀ2"ֺڣ%~h#.B^ cL]tvϊ M0~"-*«ȉ&$05ha)@1J"۵r|ʘ-7_lFӰt30]i%G||`B\=Tc"6K>lKF.zSU C|F3GVίѴ6؊dxZ 8Yg5Vԍs:8Zxg|.fglkɊV.67, @ʫEHZqO8q(Ohrv~],&zvx|ip^\ڱ:.WҬ ^>[2:/M_6b6ں㸴7mZr$Ϻ\zS.-Iw1?I ɏg8obmRiaWǓ(OYIϟ߾oo_O7oKE_ϫo|?s9˱U]';; `}][WߢkMME~v_i2ǩ5!nnpJaU#]%J=sW5{bQk .v"J xb{4c eJqXXU?){nFogvbXBR\*MXw.fTzcŜm!-y@XUİ,q- ĵ`2԰G&y@&dM`LF-P2.BPЌ`֡!:bd- A'PI7 VN!JFlQ&1ZRla5ؾj8My6ﭥ:~[rG(O0W{*a*03gk"A9E!ԲDY"RrX$KN{gŠleol+o^Ik&&']zrٽ>YNg׷hp5ZCut1;M֤Ƕ_4= omgmίm-FfMiN|u67][@R}c|www7ݛnity-al Wn/Z-ZrPd׷9r۬lD|.qƨ7$6/b[M^b1 C{ qޱOdG 5cGbTYNM(3?S +?}'YRr"h51TWٺ&.F0Bgh󐬓X=7d_c ɻݏU;|ǻr.SDʐyY?H)(E+:!ۘIf7:UUzԓA4ld)e+ *m4HXPDA<Ĺ]Qom)fipNv:õ tk†;>i{f`ń|!床"m֑sAI!1b IjoLh,aI x~ P*Z*6=9R$0m,N:ٕY3ڱ_8<;5_XhGg4F+Ҏ?yy;.t߶,mi5}eD7ඒMW{&C^B.ԙҍ=xir` خ9! Q? Kh-@l>UP)nBi*!e~*bUee~ԙQ8#a ʤ"!K #0h3&EV\IZ**T(B0rђBʾ0>t-lO'%%a p=24ذ#bR޽ɇq{qE&ireoJ Y.ĊI"L)4ʉs֬2)m{t{4@mL ʤs@I>P̀XG ZǾE&J,sPW|2^lu[N>9:,6wgOZu!%Y՚uF-C ASZOΪ†&x6H7EhpZ {( <x,|28CYc7hY-0ʴ^J&-PDVDJR'wi̓-l(2>|r04Q#QW~ #>EGD*X`᫨~*|UZcJiqx|N竸͏.\/]Q[mdke g"5-V%`-tnI G;p޵6ō$ۿ݇Y*";؉;3;'tt3 ~$hA`$#3dUIU]v~;!p9ń$$;uJDNsaD_,im'LT^16H SqC| BZ ӁRРztvV 7g_]JIJŃ\ǖ*ķQ`ooF&ww[+97ഠZP,TXrQT;C%ܓPJ4 YeQw{"AzMM!1C@R/3s 4F3'*$*R2e=MA_P؊mC%jTY ':YO:MU6֨\ Vdh 6쌸Wv[U%jwK7c! ͓F`D($k>LjOe4*!{!ˑqu7pB[w,V޸c1߽x CG {4. Ivgo.Vre1Wrkdw*sHMSUŧjf5}im8U3Z#ſ _O;_X^ځvm>PŪtBCLLkB w8*fX.O8#X/Oc'WV9 }=6F'tvk0bV.•lsiXޘ_In!_+{DV&:}kd#gBZTJ iN8GY i/p'~yNQK3'q O%`J'):UsՠۂW7Mz XF3JJ qGDL!NDDN"3F+BZ%#s+~#o9~yV3| 1IWhV:K h1jjsry(h:uEMW3y\~2hno/^ ]/9KD%J- -qNMIػB8jzq!w $ha\u$/)G\D ,=3 P,5QAt"CgQ&5.pEp>CVA!R ˻9L7;=Õ\g_}$6r9 SX.Tc;n{$ ǫτW}x&<$OFauuU*n,>n\Щ~yxGۭt@ص@ WYj߸r:B1.-Q)B2Q 438 82]R䅋KL]SݝwaDžt%JW&/uoJ FGŻVm)u{(ָq(!L7d<pc( )*(=}H( ˧oĐ{r6WXe:Dn+ 2Hgmk$ BZwG~͊#dVnT.vX]VC>PZ5v!MM! +JT2mR%E{rTIKGz&V,U?ih0윓Aw,t2{@s*3Žo>U=ipz~EfBN_)@WTm8t6WжѵWjMHO)7͟ '1E,d-tDOʃcҀ*[r)iH%pT?=qzRszIuOO~]vAB4:~+餠:qrifZu01>hOtghDZ)IAZ!LNX4.jg FnQNpǭT$HΪsZGGe:e-+w+\R!; WM-ut`ctL nB@8Eq,9r*!Y24@j4׳$O$.DV jD: TA@xB9S=,SQD9͕RMRfJIǨ.2ugUgC=+PJPGw *& -Ymdmv GeH ث@*pC/ s3S*@P~ :X [BYd*2-8SĎymgO'w&Q&;^ʸv?`$V쫁J~Fb{щFk:'kPN*֧NsqB&Ӵ% UCN"6ќ,CI&~WHHWM$k[w+"XIĿ.J(_g"򩎕o8`^;sR, g"ptдk)/QPb4CEG/v 0VծhGgv莵C@T!EF|"&0ub;W&i(KҊh(?:xkYp6Trٚ;hHh 0nOi-Si R2|T(j$SnPOQcϔt|3%Y?gE9ʏя'<@/5^۽::yN:O(X*hN>kOؿA%?nmaofhI`X u9{?Z{kAx‰x!DNuf)z}?u'/V%s+=XP_~[~asc(`@u|T|23!~xᇢ<7~h~8/P+yUEk{d~o{z5Co}4~jy1Ip uɣe8P[JTOшлqW):Ҳ͇tJ-{yQIhCQ+Q&r# QL(*a| Q<1C*&$D\Bq;"x d96vYupA8ɹw3.ggmݐ|z|18/ovӧfJғ/(N IeQ03) sN&537dFDazDyנT{|-mEBAGKj]CV~5M? Tnm1ƿ-g\vG ֝P¢vYGqM_јCY4k9 HgY2o k SRco\%c̕YRg9&ݻh˕ʔ̎ėnX‡R*%T.3yًvt NaH*ݯh/UNzetZ p yi$>`uJ;h|rIir2<~ nGix6;(dP\ej}{=-RpåRuVdhh4I3Z$f5XT:!44@p*BAo"dY5$ax!Dw^B S3'dC7@ѷ25뽝\pXoEc3Q.GuV,qkeݲӬ`XU2:k"7F-*|e2-.z$7<dvzjzz,O \KDTQ#ω)rV'4 J錇SITZ*rA?:}1̃?i2>-sǧA>r4^*N$kC&FNWZDMs`ƵffjJ(Á%4o>|;i_8&]5P1LeF좕zej^]$ٱ%y8sjo04x:8jGcgu|p㖐@Q+slGaXؤ:VՎ_l},|b2#\teny)n56HӼ(o- ܜ ZkN x:.Ή}N̈>^?c&>~??{oO&CJΓ.P*|G?NƳ|=t6AylrpǓ(̯/S f]rlѕMOdSskWǛ_ga.14"U?=]YOy,k}nF ˟7J]m6f}qi7S։"WvpH %D%Oű t7~&{jjoOex9gG#!ۑ<,fW8yZɗPA_?(x[+Jgk{{x32Yw企T/Tk%)oD-͢KJN'IZxZWG_58y_Nt73N!y\ޞ1Z:ˁ%y[3)Rn}W]d0yڎA ^Wdbmc2̟y՝qCƴ eh]R֦BxLy>̤o'<FB[-}.9L0_ix!yN:Oܺ5"(jKg&ƼMY|Mfdlۛ /]TdbOP&@'kQ&[F7_ՄQc 籐$d HjRxRPڄDL5$}NG}l^;hJZ:A9́rVt2+RsnxLOy(i>*#urkHß3&(/|tTȬuˉ^幰]_V -?ZƻzAowkГ^oiȵgn~6.5yl8]Mv;/D .݂Zmn[6ڥmibz0 ;e/U{H'DWgB@]2E*R*ቩtJL2s:v^OXpsZ6$ DϥVj(U5&3A<"$74OZe}{0k}{gB-~|fh5ֆv;{"g.&䨚CXK޹xY^M%†<B(*ԺU%p)*fhy6%GW^c)_x\:yp3 &$cP/%p.%[(oyL BUOn̩5v;]|}S$t@m޹ϗr47SV%2'7*-Cmk֥qY7T&"Ĩh(UXj#ނ5Ex)Z5c+n'ml8x8u{ϖ`w&A4˧, jQ|2ߜkm ɹJ*YeXEٔ y R ɏNGBP! HC5?:QQab>rPey :Fd`X焖&Yf QDwJj.B>T2XグBsHFbP<ņK=Z.Ah [Y΢پYq b)熱.]|3]0c6UJX&1 Z5Z~8k:*85Ǿz^j: 7Oxov*ȅ:1~? /rWJhG;x( I]pY~KƄޅ~٫3|?/]oxo6MS~?,)ޭ{lW,Yu)V?mfЛuz^v>|`P4>3&3_O?f3%{=l{7JgQػ^C(ǽ׫?_zF+^tBS(DǕ!* TR'ӞK{G'`\DSQ r,:b#(A;h->(+P)G"81͢0j'CSs*_p+YS)y‘9iŨ>""h֜Maa$0RCLIaBA9)J'qSVKDϓn&x:'t!`uY -ytI҈  BE̒df!JHȑЎ)MX0Z•,g F Jȉ50Ύ|Vf=oK$AH@J]RmI p19-I d9!ӈ%&v,P%y7z9.$hz  w-ezf-4(^R0uD!L4G1FpF D4>]XFH\rLj!9S2Gs)Ʋbn,C8`(2[lp.dyE I׃d,mo|E'|i( r!b$HsRD̕sI\FNm(MǼEAo/wMm]n BV=g .i:+m<~65!WS9k4=/ 3(҂T"P^";eMu4rGs[6CzNߠj ^`XGbhCl,{6%6jP'gH[oJǔ?wq4)186qv yǯMS[Cۻ '/_?>N/a~}wՓ9-_c4k?ռ?s4n?eSCR.GӛgJ۸_җ8HUrgěw?T8%yGb%mp(KDZć8 Ѓ~7o5_,'W) cgxvSp:ݫ],Ƨ NY~4"JgzLkgyQ#Lu~j0 NwRtm[^~F[U/?+UHkgkS~ӼK T5ě9MN [z~u 3(N5N bS--.Xk)(gp`47W?yy%t Bvn;'N_HCe}ژؓYe CJi+P"lݦَw,SSrČf|cf~f 7 }l;m`8huv{g w,wR&ي ڬBj:6~;k\W]ݮZ3P<u'&6s"Ǿz mg5O/<:K4 U0BJnXk!F ښhT˄fv?kū//^5SFg<]d"R2 tZpV\LGwMX![zI](@6ݸWFg^3{%=Vmy;RJq;oCt(.q΍E$ ?-Αr63DɈr x6vfబ]%*|<TJ+VxŘFZı$p3*AqZm+:x0!S&aXʋ5(K)µEzq-H>J.q<^?}"(mɊr3 ϪEpL9eFJg8 f ^ 6g )+ lS$0$0m73ިF")ӱ2&(U,*͸Jb,"CbƓH&6SsR(P܋OI}0m_RD Yf[ 1H[$+pVkkdZeh[_]폵4mJQ9#9搦F  >, Fi$pLî5tv>z)"<KFGFb8{b,rݛAוn(vυ=ďQJ>%"}B$Ou&E(/l15J04Œ om4WݧolRrx,f; @ڞB# _HA)،ǡgW `8tu9 nE-6`M@ZX.Slt[\ɔ`B:ڷ_'= IŇ00>{Ww1\C]YLzÓŏJ%H#Wm!7F (:0:v*aCqfyf+o n.L|csqY,3Qq:j4vC%ChHGfA˗YT `ݫfcvzɺSzUW&ӫXHuȥc@X|Anv!$Vn~ŏOإwW[jOXf%JРKY+#.Z1!*\J.#mv s\"*y;P$N;{qM'q#rF\W9RK7baҚZA]$G]z* RJ/mPqxP[Li(1#Q9gbnjzF| - 6 ~*0-?8[ds^/0S+dY֙ǬtM9 Xd;|+8W+;zzS;e(8Oy^KEspQN95{YiAQˣ:;5yy>ykgS2zՃ\wP?$QxGߟ!Rnm‡&%SjSj3Q,R +;ֻėVi[uU3|ڷ4L]:kz[ݶ&p$ϧM\{Y)IK M}lpշ_VEi';\-ed:k=/:D&<qX\ {'f9{ZFDyMpԟH8,0O!e\P\Kp/mֽl=u/S*4AKR"On52'̪T(I_A4o6 AHRVa SX1c2bl5ͭe6qKp,U4"گγiFW 6nꌛ"=ڍ[t!gJ139evC׳UxI,;S1s钐kE7%ʡIp)?8]L>DM~Ժ[gn]Օy+ƹ]vq M -h0X_sus6ܙo~.zSꆎ ,yS\/j646zu m=7 E_dn-7"nsskhQ׉FHRQbJ/mdǠ|Kٟ)eF ~/YSvLRlN2#.2cрBr BD߅cЖ+ {;>伖AYlNB !12#yKrJtH.R3gp>qPyl8dPTq8`$PNbXXTJ\8\l6ԺNfYj-;lzrYo6 G 3T|z۱G+#lLp))kf=,05{KS^H(" ,G @Op).}a|J#32x5sf"Tnd&fddlq_,X~4y+.܀_ݓ_7(M;Op6Z2ӎF,h}uV֛"񖖻>ws,$Ѹ4]R lKnӢ$JђdKk,)Q?-m閭>$~SGKo<:Zɵ0ԥy\ Ղ"y*^JP)N %vm, QhR[Z{4sE50ߗGFh(݋`f4ŇE6t]OA])n?6 dUǪYysU?౑>@|,&m s\MW)|Z8+9 m-ڜ=lRdT%G(')~D?8z< Z>m(nH'v:.nlO&OvҹK]nEz}L$K?aM)-ٹjJɼj LI|щ Yl{ѕ'@yE*H W".gWJhu%OHW&ntzѕJ@f͸D]%O߶,y(vRK=m PXz,I'N;R⫤!x`Gݬ;{gOې\Ya+λ?v;_4l(Iy7z,cdvV}0}½)\Hkl>'6ǀ<28ӼIOicITFI3[C'yqMZ60#])nt\6Jirn7fW]\I?o.^t+ iu壥`Lҍ7t])m}vݲ+=؎tG~t/z YRZJ)Z8{pԍѕ҆8w])sWKU) EoJqEWJwl<_WJyunʤlC?RMvJ)y .QW"HW])-~0ZGޙ\m%wfwx۹ΫDovAM(7bބ4f+m״RuǟGr`Ջnڽc7]'ۂ.ZwttL%N#$ ӏ[cNw?ȧцF̼RE*ʚ*>m>t`:3FEWeCsוRjr) nt~f4Z1sוR&j)EӓX]WLEWJKnR@+6%Rq<]Xw+ŕnt$b]箖_O^'~g 5l7HigNn{xAz8>/GAzǏOL3xD$`c7r7`J35Lfj ԢI#]8I?]Fe2w+޻ؑXY\g^i핒y%*@i{(.ۋٹJ)]-RW,Vx5D֜ I{HӁNFڙTmaQVM5)LeM5%6ݮ9ǣd2tS/Jk͉4֝hQ4J`ӳ:Jq]7RZrn&z]9GQ#]X"u+&zѕF7w]9Ҫ;3ء0 q?A=9œh}J)iͮ+6xHWu+u])-J)"uŞp4$|7R\ߍxRJ#;ۑ83T\fוR5Z"h{Z`1T\f]i%]WLkvH] Q+!6+ťЋ6uɯZŗxrt`q s/RZ6sוRjb ye%.LM}a&L3V>+mUW UƼh\])p v+%;w])e04FWM 1z3;J\$"i'D\fJi]WJ)Vig¶R.lt1&$1Ȧޱ4 Xir]'Ҵ=ꭞӓvPqR #fСJ2;h:9kZce֚qRST#ބӯӝlk2םx4Zo(if+I4ϮiSJ0d:ҕԍt Ji+Yj9r"ҕXɮLEWJus[Et .t+JqEWJuhҪꊒh;p7RSӝF{&RjBԑX|7RSӝD]WJj ) ٛnt͝A]2u%ƹl+;ZȠ.+e2UW U2ND`GWL/R+u!"u,tf Ytw;ܘ ̧^o6 7x4x(efυt4VM?9%w-Wt;8Z7s ϔ>nCq {J|M¥SFFKf"(Ը5U>⧝RL9bǾR}k~<Օ&ƌmw~x??@t Ec@ݿ迶~{P'ڕξנ%ۊ?uw\?Z1~}}{Wۻ c}1{|m_`oY]wtw'wɋK`b>kҏK>oއ=!j^>4gW⇚ھ5?A_U߱Ve<,,gt"H eO;g@O 9sߨOO{7(#"{z^w8Amn+ t:;%HF{26 %Q4[|5$Pm_D{i0wڛM^;u}|;>Rn̶DMMєٻ!TBa {!;pn2\Ʌ4oT%zo`зa12:x7zp0!Ͽ@Yw-;B5_O i6a%,pgGSl, e΄ R ǤNPQ5P!qy,9X\KKl457"K\|#kM܀ldR"%nr3fL myhҷh̨ѹ`1G!R8ǕiZr6!1ϞfJ_mSd;DOc`# :h<9hrøl)#%Ro RjCEU8JkcƘ e\ i *D!DnN&&]Xd#QrGIlSAX2AJTCX)B:ܛ{jie\s#LCWRPfI$7&[3Jg |몣IÚ,9Ed,-q͒ѭw3sh>Bj>Z=]L|LXJХD@B "KBJR8EB"!¢=`QkBfYl5$gF99/yqHܜ=YbQdB.Gѡ q =7!e(vD`F֙_MH+L!LyLX}Agآ!J m-Av*9|]m!(L260<*0dPK^#!q2Bmk4>Z!nS֣@3.\ZBkhLަ1Hnp=PP ^M*1*AUa5)7fzaK1zWmh@*"JÈ{m@E9j )&DŽq- בo|C:D%+7d\ * ܰ!^I6D,FPTAWߌvW F? c,V2: H}K2!]cm@(v HՐa%dܐc᰷e10DŽH$D%Bi$FWg-E ƭ! 6 60\&#!Gr 1SQ5ib( Sq{=O _נ6 !h{ T\ D]SFB^Y2NRt!}aXgAPS*Fy; ]qFGTGͺTHo[BcAJ~jtBRlݛ"n@izYBB $B:nݰKNZCŨc#%3%DNz[ǫXf0IgFJ1f(RjTMМ@NHh350vs6]&UX1+}*Z=Z~d? &X6$X - 2:d~H7;tV2OT"R0JZtTVac y$ ~X|Qa1. = q-4|NT 2ѧ U2 ')ZChORD:0x #._q) U/<:ފwD q ?,Fu,ҩ3 *%*F.ۆP+a\*!H bg@umn+|v''kmHЗMw~Nw .Nf7@X\SBR _z-y {Nuשyml6>D+[0 1w@1pi/zs 6ԥZtt5.0_(!  D(Iʀvw %txVh=c;&ETCtЫ PFM @ۄUY{Ōhoz:J!rŵAh]X zhlXc?cp2+[$aԵ7nBÚI`td@mt*LykmbpxΘVE9kqDs`4vƸAϷ? `y05^mAE0:y9\n8М]zD뢄>k,wᩀz`BRDt^Z nRnj@CX_Wa]q5B?o +J{R,r5SXVkM'Z.r93Ɇ{%&]̈@$mkO02x÷l-۷ eD|՗7P&q$fA`5"m@vaq.X-,ae%B [xz^s!֞`Jn { ̚L=B$`YPBʘ'̭&^B0@\DmhkuݫCUQvuED,0貍!;f`f%dSD܄$Ht!Ē};L\"K֘p^{n銄 S!֌ P GEWI~9n&\0̻T67,R5 V)U}Ml*&2N9@pz9sJ UTYQ*b0ֺ?㗃_`KGRt vIS`"I;Ѩdtd~ pI)TÔp @JD̲-g#.!?46N_q6+-G~asI2?|o~Ԧnr| c[.| 1va)|7Fl%O ~`UX'EJ t@snOF286S"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b=[&o?!&P>&P׹[afp&PR:b=G&,p"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b=_&Ll>&@RzL 8=S@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 |@eI0 z:L H[G@ϒ `41 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@χ t=z# ^Kiu^7 /`14qr>?$S".A  qO}~v䊈Kρt ^U'c "4\)%s ͕s +G'c}="~]͕,34W12W%a=`*d"ǎnNzJk+:BBM>d\@p %r{ r*tgUur0xy;xqȻY#Vo&W*8^}Ոv[DmYu/gT^ۮ6tv>]FJ_OGP]OB[tbr>J?|qmpM+u&_9rptp7iNRQi^ȭԩ3RaX~juzjKORh>_m[{'0<=΋S2!_̀b hiO/_Yׇ0ia 1c%mg_,:;|ϖ[ZZis-^ ~Jjl-CiU^]&0!y0h~U| tf`q}tqb\K>\\w~&֟/YĕOhBZ45wۇv8^4R)Cũ&l= >Eg[rcA%jlX 1,NQ\u;U.y[[ `ϡf-kqٺ1ŗjWdB6JT~~j+ cNCrӍ%w.͊/no7?=Mmy,Zp:&,ݵ\^4:d!S ZI&UHV#fL\*k.QxHJfnYss> [ž%bGdk-UHqx0Xs%/.}nBo,j`9| \H{R(fpiD2_m+ߵ>,>x ή Saʕ0p܌Nn=$P/Nx5jsF c̬ma0M۬oߟz_SxEkK>3q*컯 0GiS< ⢉lkJC!A`駿zoo~\ط/ニ/@k]ϋf}u]K9ńѯ}yK/Zi1 o7\o14ی1vHĝ$$~,3ڕG@,JpA6La'km=:IrP8o8,Kծ?MXLsl5 ڕ7C.Cީ\oóYc<\\醛؄9wARc gλRD#Zs+w Nhr3opu7iک昗P:9/ڻc{N2π7v:IZvR:T^UYuq`6nU*Ȁ>j=bֺF|(KtJcPKJw\iJT[[)tU(/,oɶ#eg>ʞ "{flܜoR HW}:zP&σGvӌ?,%4KB46hr%[mr(uڢfjkX{I}'&8Az4X>ߌ.J\P+_?KÆ[ ?n^^sZ0'avi7f閜8ٻ5}bF|b w٬O~&m+v }&tJQR_֨ MR2,JĪRhҾnDn !119aǖQ6+aa0zk#eɦ}SP`HF*c27I+k!tֵSWtzDs;:.YG9?<,o2U%SC/gO z}MG2['ԺA1+-?*!N]l*sm*%Y\r&&^$xH᥊卷N\^k3~ I7IRwcF:?0s.S𲏷/{g6'|z&!mI0E'ged-T̖OV/_ߣ U,n^x'gM)]\[ȝ&}nO3_Kb}wW?g|aJQlbsF=ҏԅ~:Ϭ]'B@VT* _F*4*VYm=^scAr_ 8z-scQi&3NFGU/{BC-&1Eyb:=!sfvum8~-FRJ6+И$^k}T@lrVwHIciזJjg0Ŧ 5-qdUa0 `, _ \dn^|bz}ESj0y"usUFuuQXq%pPpu싇~0 l`yQ2LJ7~ RE !XjCB\CH!#=XbD#$t: "ϧWC,U씭ڔɻZ|ed q~mîF-2/ RхqWHm&8S7>lgOv|9z. J֊xi"8PtaD$_.}7E@SwC( |8)ֲg 硫rI1ҔqL"fJrSrCmGxl=C % b=!%X \ R&L1J tÖub,C~=cj9Cи@<} 4iM$KƦ.%61ZwPSSL6`1$f@5&o(lCNJ5vsx- Y^]6W,>e[}(vQ9W'ΐjCux \~2i}7S @xkWvrL{#iicSB]o_B8}| ۺC=4(H_$ŏ }gyy7}΁Ӫ<]/Վ6}צּtpwso_pu)}=GYHx+ۚ\u};B?>|n}+o>Z|Z\|[h(k&o7-Gw/>[T$핇;Bwg_s{uc7vfZoĴ8d͙:޲cֺwZ`X>XJ]FuD”8!wLr#{PmR'3{#'sb>UO]$2I>.8^6DLujȓLSjE0jax y `QZ .f>E/HHŷjU^ \Ѻ17=PHUo:^įi=smO8\'.4늏M*ҚDפ`97>HD -+YEpK !` !X-E_ch|Fj4"T-h5ay܃w} yUYq<!i1ōVb,* ~SooPdbۏg'&?/G.& '%YQYQvW g sir!%އFF6Zd}r'k\>$ʹ$ϗp~ opX.zsܥG~462!4>LccM37'!+[ur~j)x}R'JL85$'/ —T}5Mk҇R/[ G\ޮ^NkfIDJIL+."$Hrt;2;P9 S*giU,76rX|mB)^wV9QH%\ffϞ;sswƃg?;i)LO"RDe6}V!IƉk l ~E+hQYQCƓ$t* oVΥj!;φ-~߁ C/HdK 3{UQf5u 134kMT"d Fgo-Xf)P4Rpŗr32Gkuw jZd%:Z|8UWgw.xIHMq7:c1JJ9J0qWp >3p erLyMUlZlyhϗ+A5ClRSi6媥?Nчd5tKκ&l%)M*feDtm_G<f": v[M'eJ 99[.Oћdf#Πԧ{nNt;26,d_I/˾x~6 dZ]2/Jv3VD$̵qoSu5.ۀ”PMNijXq" e 4e5 E~އ-E8ӭxl<*`}T';gԼfJgJF Eнl?*UU̞<8?QvJ j2SP^CϪ}<ڨ3&f6>z$B%SJ2-Z%X4}U s dh5r)$A1/#s/E" Ec%60Ւ-Hz 88Y6l9'H}2/n =GJ_+q?O^o~=eT[!v&GKT+9a^>Jc?ilZòCwރ^ݬ;j6/y$;p8ls|F YZqrk#i@<㬭̹LN&*NjmrpvN1g$oǰ~:g0>Ie 1Am ֔! Q%`b*<%&6CpR+%.?fl`7KE0|q/7 8kSl AA^@IRPZFEun`C)-jؘ>:߻ʡb -QYM?rT=̉!E*iFّWץ2|Σ۱CxǬq;VNuRG v91 p1rWcĊɱLN- [?HM-=jSDM_e+7y;>O#C'C-ёYz|za`h{#OC~0$2t`D[J)3f[4 æB DVWh4gFWZU9!}"d6Y7Rs%Jd]才_ NjE\'-닫Ÿf t+WpUyo)}Jz$rm%Ƶ7?\x\S$Ç+}(٤( ($f0ƣ1>ߪWd⅖ǯA6>7hml0n|~9'ׂY+KىTj4rҪG5KW5S|f~ab{YNjyb0i`\Ƅ7Y?`>( M2_up4CgaXilp6;~k*m놉~au-zׁcxmx5#?lFФ0jX$R;v+ѕX7 f=r&XMh\w{ME eEi*Zw'5 y0Oƺ\5?knot26VKvuM$dс2L }}n[hC7n͛OO?}PYj5ԡZR"|x>v\Ogfl>[c_.Axzvn0eZc=90l'|Tk2vlT|2gIgbEpbno`wE~ aY`xlŊ,?/avo+K/ZG}TcZW/,̓KR,.y5)nU>ov- USq^UszRT/UbtLZ p m0t1]^o%=z-Y ?_v1J>}s"ٖ}ic~`O: CJi (&l[^d45C̆fv"tvЕ7FZf-'d:xuv, R>ٚJuݤ2L6 YzKk7@tg͒}Ee  GmjG]"Фp P! `${Ю՜˶'gny{aj߀_b7FV~G gQ~դ ѬSxAݩXLMjQ*({`[wr\l.ƋRԔpyZuߏ}f2m Ʈ)2MYGgՓxnm!l;`BSDAO'xЛ)^{VȜzL4P |Z1MmH{[_`vR tX0MoӴCZ|0lp5^Kokl?Zl@xv7?&.1!}N oGtt<D]7-6b04<_.F.yT H.^?S:_lK~IQM)53ZX}+#;,#ed4ۨr8UHZfV`HkO:Pby `yI\!Un>AOfG]'d=&͞LwKE&y0ĕ.HuywMRU.!.#xv{98QluEJT ;_a+|vW ;_a+|U/e-bvW ;_a+ Wa+|Ca+| ;_a+|)|vW ;_a+|vW ;_a+|vW ;_a+|vW ;_a+|vW ;_W ;_a+y:6P=O(O:6]sr0oHkZ̈́ <ֈٔo#XN< >He2x_OknsjMPTx ڧj!)DIM%(1Jy-a+jpѬF(V:a88dv ( e5BI>,3B;&FhЃh- kғWÚo'ovWj^yj%"kOnߓlWeE[xB&D&Rg{%Vyeڂc@PNɹP(85/Os.C6 .zvE:Iy#q%(c@cOv00{Рh$a"DHD"# FT8G"W BJ(2fֳllg^ P~wDM?]J,G`"ŹUcm$,)LkJ, (CD@i # T49z s鑗 P7E̴RX4 Lh2PNra񓊟TćS*N-p0Қ"2i8V W# pԧlsbO{Y"秽e}w2oY4O݉QRc4/nVPaa s"#ثDe(ږoGd[LCbI`g|٦Ȯ'3+_dn `NYRjo @Q8hʺEI90hmPD0̨KcvšhGvڡM,- OO+"uSd(ϥkPC" %䞂MxfmpNxoƘc|: )Ku$Ԍ"_kY. 6J~>b6tkv.TB9߁$CCwÔ C‰ɂ Gjd~sl1b1Vٵm3']M@em.x1 P^T}xFʥIt\UWNnpvl^ލFkl$s\gA7g?Иuy(w%wڗM? 3 / Z1ځ_oe®q)-&T Uv|1'Z:GPlꀹLRup,&Gak2롃c#,R6e4AKR"_֧(T3bRIdr˺yOLG e(Z D佖`Ԁ'R Fll+b!(h>h,5[m3XnNC[TxCF6F"ƍdJv UwU'ε|p,ׄ(Z8IWOL2n>r]ۜTՑJm錙ZK7?rc߮V{?plE}+%7χx͙]yŜv{K*boxtY}VZa95E^8충j{npN?:$}s+=)LyoRgiKA9I&vaN5|ME+?٫EH9L^M`.1()3b&Ba|Nj,+mM7)ۜ~RܕMhDAYv\gcNѣs \ >9 ^חm?N^vˣ~w]r`CrٮӪ<O `Կn5in3t4k^P=yjԤU`#. fcu1^LNӪ~CO B|[Ʒ {YM~cEbq,SH}m#s55F Wxp:ƝdɿÈ #Y U$^ji:2"9 8h.$w0/4N1Lha̽øig-? W.'Xt ̛S\l)4F#Ҫ zTVE I XD|0a%r"Jˬ"hmC6Ozº_9{wKo>NX# i#lL< s V%X3 Sù;uj# %ZE@H1 N9ť 7 ,3r#c6qnF|J6,63vB I½j,?2^s͇p80_8b-j`iG#!2.x0͔HKP RE#nH^JD%$ؤB:$R:":0T;87#y0)s͎]QeFmVPX7SsF~T\;Yi;F# S$=/Y1h!Ov4H$NQp=řĹ'\Hx0*r͏]gFD^ gGIc FEne0/`@%69@Za%X(QHLd$&pA94a#ֱ̈MCmg}AI5KvEEł9T=ǘ1bc)A iuH.>pP0Uf<1m ,iW)X1!]^!M0fjɴB0N98D|vTIkKb^N8EB'$!BtK4ZIH!k dNą(4HjY0ب1",DD0i5eDDSA #(H8Hge WO?.(sN N`>isk-V}桸//'^2j R"/FK$4)cg>2fz|Vp1KaU(s۔q.>jHUucĄLػo)=\^I[g.b_D0a'jFmN9wmI_e~TCq9K ,pO0E)$9~Ç%JkXfꯪhzb{R/NQ,)zb5ཊ6GYY3Ӟ!^$䜇$s!I/m%S>H^5w)q!Xi@zcxgRbR1 QWKzvzkX'2$d.9c0880(q \*g 3/7Lp_1-Ӆsm٦nz6^R* Qj:kS$k4fDr49ehP{{!CcPifwO+uג |KDsi:XY4ÈhZt{ x;QɗQ JD"2' tgigM2Ov'ңKF|n R̉ eI&i$ b1z@FI1q㴘#*Dg~ 9J+.M++1YAh0[iur'%տyǻT JgR/7 qZVRrdNdJf m=(n`=w]*EPM]\ ,ō_h@ᨉѴϹbQĨ^L_sט(ȾY6^T[%g+t6y\|mhi@/h8)'"GP+=?]/5 S2}sۋUpvS-9u<FVˑ] ~6oOh94zkIږoot]3b}3+Z~UUyO0Ych(l4eUnuuR&%ğ34>=^*Wg=_cz iVՏP=,Xp A_Ze)*Jvw:]ynU}nRwOFX__Pc1P#O(-k댩>W{\gNB^IyTc*g략gͥx D 2rDuYp M)hŸBƲkpSQXquGL>cz5Xϙ=Nsk/nxأ#BnG )qaAr9!Xkb Ym5:Eʚq={=e8qPlL2 R("3ZRV rD@TB`BeA[  )i],#* \e `{n1! 5d,mkM!Yx w2%湡5'`uP z8obbZΞ>78i2ݐƅ: Gxja1_{E2*0DsEsGE\Ĩd 5Zd(0QcR[%XBHfb5g=jW]Bu{dq]4|;xΕ@?QE>>*vЇi8)k @[Ǐ49A\C2Go,T^(z4X4{xt(NLOIu1j:VV>O:Y|傱 @kK6K $P+e28˓6BL4sX ز+5Z=֕ '.m9NI.i1T *O'iTM鹲P~5!&頪~}뱍u~n~}}ɧ*k>M^r$On0u-X<פg(kW#jAɵFrrj RWBb3U͇mutz;z\[ԭCSNhΫ]].Ǔ8(hQՐlD}[͒d=7YM=d $S.`lp lm&KVﳊ'e k_mmXK]7tE6m#t5i.]LOflTaO|kQ sZZWNY;uV"sQeÍN7h +V.v1{2!V%˹q$8,g>qM&'ǵ&UmĹߊy<>;ugM:;rn)f9޳ܑg)ae0M` fq/r/IlHaaF ^5?߼.y]M_d]vr:Ig2qN\8?v*.Ii.U`4/$Yi)Ђ̳f!"~p沽fWх&N2}"H.Ocm3tU_n)l"H/g㥨BPsūWJ@끰9ֳ@/pFKJC 2|BiI.\ʦ>D=gr"^{͋9+*2 >eH"dDLD=hJ9HmvFnL0HN''g4r 6pPU^H0rERT2͓ۖ8:}q:cH+Vdqx>~d"9nн݊-C ӻ;r"mCD 8= ONmlԲBЩoyeruh'2Oo;%Ă-z) 4<'_LDDnG|fJq>S?JWmY:;UihB3ZJd8H`*lczl4lR Ek91%c@"t) vY[܏φeO =%Wg(iRBEҦL @9D)\qބ L#&E+b EP@,q\^1QF%T4V.D[3܂PqJԐCm|&Ph7Fhkā͌0:;Gx.c-@R>z^l9bF20ܐ~HMB|&lB kkL̄4&e<#sl8z̵!-hkItr{0(eɪŁ;ezÌ48Mъ6"Aj)YUku9䆯UEZ~ZW6埌t9Ck+(ޮ8ݬ<'jCLVOۯog٣FsH/g:{ԁeY~Æ ]LæQjFџ}^f5TRYыTR -[uLgMy&|Яw7$Jss;%wMлd3J|SM>Nb6~7lWV-o>9/F%nHJ0I yFf0F|e΋eU`l_w¼c{LC[tmݳEJn^UNf:O38)筒LG E24AX@kGb9.i],#zʐ YxoI7d2zZ~Vho8 H8 vΎ\[+w I(w[+{oͽ} U0X0\݃ zS=5AOMS=j i!'DZ)$5ڤU*uS:RK%5)H ѡV0E cIz!&9s,˸Bi2knjwm_aSeHsJʵɧݔ璸H(_RCACǯ{zz>j&-=Y^~<ɋ[kքv;Cuyt'l^HiB4j8LӨTG3]>!]Nvr;O´^\F{qzqy> [pI=B(`*D0\¥DN:tKpH)e;$/<6 C~Uhi?Kz*d' <k=&IFg"z*AsI-Kx,&A)L DL"{mP*E0$O^DLjl0bvMlx#QDwJn.B>t2uA@O-@k|LS9W90lgiK7;a/ltro.xnҍ`l+0of1^=^5FJc~4l}2<x ڤhˢԻoߞ0.KojM&ȱeZҚ*0^]vF`s\(:E`R<$2bVDpH.)W!$/3=wabKKV΅9 C ȶ\.GA,5׋YǺ4s<~:_XMAXNzf sqjO'x;Ѳ<䜭wvw St$%7HCĸ t\y2Nvs7H%hs!+ AxmJR"J`$h۹m}$F}PGLGvUa5Ҝ3%}F'a2.}\Q.B|\zW,)<鮻~LgF2@H&A#3xB"v| b!jBxrמ%!Р[05)c$&(`AQȤ4KhYVZ#CIH [ yJ"I-yPpBJxgnbB!yМc 6SSc0m0aF'tc$ r;UE.kz\2?s I,NRzRP0 0n4 e4N*ceX+e3N]X8kVnu)*-pऎ;nԠNT_ hO$(,P0>0>0 ܉" ȉhWySKBAICLVxd>Q]hBv"k(>"8*=E ,/kN၂acU K!W{ Ɛ6,7(L4q & %Zyth-<)T ^Z(\dB21S8G!ߠd$Q >1|1?vGDHާc,0 r VN*eL #h<_+/'Q#ѽZOеwjt> ,%ʀnZQ[|5j})͹y 9.)<R[9)D1-ǃB-wH,Bp (A=!edӨ>8 Q1˭@SV[+X{)&,|cU֘RR ]ӕW\ P*&)8e 3(%LߢRs諟@_lZQE)xIeܨh2hrٝz4Bjo6ٲ):{ .y:+mȿQs9ۋr9?.[gQJ )D0"01)*\~*^d wmW:S^[p$vcі.8 9@#B[%!TAw >N!1ՓeA| n3 ?vSƗ|fIʦ/xWz>9'=:xPhWF{gLjf#)'%Ŗ+#4^4B@1N1|6!Ęw^L)GQ2;G SB%Kd}E4 H: Z3kqx7L:֜uM[C-aAȟi":R;PVyPICF&!LѤ4O+'uqz`>R5D&PȱA废>y1!:}_dUoC]Oc+#Y1qP4e'6L/㇜(celox=yx0_=9@>ffD7秗a__nf]nLGE_"~y}x:9ofW|y2us9n{w{&?ǫ g 2Gp.FѠA9:苾/ƹŻQ=?c򵿡a7. Yo Pz{?.57(svgŪwxNjѹ?x蒒F8a4fsTNτ ¾ae (KnSb͂SuO٩_晞Xy63fz8ww2{Lhs#¬ z p=Xz}j Dަ[3S` TuORMnQBj65os= L @u > [:6q[d&fRuy ~aœy #̜h#ޭnܴ9۫ym76j!lTUht\Sj-DUL ApBYaAvVq'G*?|Ֆ6ZH< qW4M%%Oj-tb [%19c)oA\R|2t^LrŏqkS1cTbƋv*".⳼/⽻ h*0/u֫sFo<`p}_0޸Ld|oЛE*K#%T{=l{XEػ] Rs?e꿻Oژ pBShHW"/B)V$SAJiJȕc+C-!^IUpH!W$hxkg>'vON{e94PT)9E,!1͢T)sN=} P[l=l+s}'|_P8눎ںReVjSdRF9PSNXS.P-2$z$==4Ip'gI%$.D@-j#S$NRGFtM FLBE$q("#GjvLi’`%\yϒ4gZ{8_!ދHcuup /)sM I%V߷zD-FH,qf8]SuTwuU6)iO9++Y5q6֮ ltj;R5{dA44pNLXiHk,9&3C'&h j̦j;ȶ9Po_X0Q"Ȣf h!s4:X%bp\$uA`gժFtI'5yl>;"Gdp_!hmT耣 pT$B5ص[P&ô\fҝ^ uJ*O*4'>g>p(Gk=InyhTE"pĠC.XTKfvuD[~ZۛR<=ط{ow^~}Foq\s@GüWqtl4\ɣVuj.q'li~_9:}s|/se&;;Zq%K} yɝ4 eʤl  ēE1XQY^w ٹ]5Kr?*w&Dޖ2^6+>裱 3(MeV[>'SP0msEmHbr"h!Q\x9 8D%%{f.&Vӓ}OW}ڰ޴٧O@>ÎuhL*Z̸cAaQ[,ntQMְ%)s{f `b\`A3 ]&gp:Eŕt#L U02*հf< ee,>y|05+g'7\Nt0Y 7'ӯ@\8 2pJR$jf,22 "cIdVݕFfn(^H"RؔJQ u&ߎV@*hS; 8# ㊋3)kicQ*6t1{'\0xL/XsmFι,0xT:4-➜^WwCnF]9Y!L&&HEԉ0ߘ1@HxXMu( 0VZXDʈ"vxo$!B1yLFS&"Ij9m ӊμ Bp6+e rOh ȵL&h4 |ʈXM^'\,"vjv--y,.ʸ:\pޚ! Ƒ '7%ɂ*i$s HiԆPu+xXM;2>2/@XX PZdR+qK=yzn 6jBq@HI&H hȍ97Fh uO)`!ع@Hء@H؍51)p€^ۈ!ItF(e6Y/6 4`%7>&Z BeS%LD3ef+3jl9.4z1?h>׿h7}c;kX]DM6 Y'+I 3̳g/)?Vù eY!EϷu .&%LrcWQ_p_r},^t…U)q"OB"i:R0Bxc8PL:X:;7mh즣rVIh2 UCY"Y60X9dɑ!$rdhP{;!ܧcPifw5cDdaqk22f A[N{]P.*09'eadL;c+ģ.$w=‡& c8)3?{p eZ 8 XT+) n2Ig~py8 mbVsKޥu`ڤ`RxI'QNwab=?9wZ(ŧZ,:>>jՒ3XY: \D .~׃I!/hx_vDOT7=2Fi_fg޴'^^LZ%gQ@R3?̃YJ{ճ+$h/ O_we0Y;[u͈ͬhfVU-aLQQԃH>.=j֟pƮmU[]J&e?-,4>ߍơbB{e&hp17mi~?,񦽭z?^||9'{\.۹2}$+PTx{qh|e7oq$QioGg{oo߼.}y}s.ɫzuK Fe WDl%-q4hZؤi]AZ|5~w]ĕB!8.uYs+#]']MOơ̔ !H^X/3Cbј]gn1yp:I}a.Mk=Ҥ"1\#ܦl5f{;?gx8yw%CԮ-ر/Ď*u)w^tl\*I K q,o QlZPp`\.yqy&o/2uEFiD b~oѴ%I)$VKvJ$J+=akgxwM%O|-:8qod|W@h$|rHkg:笓"b-YVhw\ǽfoh'6zf?X4xQYrb<l7,B.е7'jH{R~Ye!ëo? Xܧ%kJ"؛%"-]"%Vzu-&)@^y%?zm"dr0qi-64pz}8BpV$PTiJ*">f90(&;ev,Gd%Qa u]LYz7~qf~I/g1k' ڭ̓ -vB̫eC,AmNu փ褴9N =a5W,kM>|3BsJ{+UM␛2m [dR_gRwdfkD 4 oVN{h刺ڨI[h@ش1anWWӟlY|Z"' ޕ>m,OV }ʵr}S,(!);+H % !q,`{~Ӈ'g?'w͋`U-L3Ї#n>Ր8) ;;%M;D@,TFJGS^!]C]]@qSwr޹ ӌ:ldU磌J@S[C{No╺>M}N_oqɜǥJb7k@|:e!EN:ի!Z]/SE_QE(n:)t o >|^Վjng7d҂ $MЮ^l[u$%`HkQ]|-zV6c=yU-=$iM\R+eB\%b;D()Q)y *ɭ5bzm{0R^VϓUAf>_o =ehwzGzJ!jzQRg}o׎~qa׀<B( U!R.͜$rbCݥg5.Gw6.ö]E Nxc=&IFg"Bh¹%mq>gDܾfwќGmzBe9}(jr>3[{/]m[xgAs .4S\J T2X 9JhKِvRTY "` "uxfR5c+(TYk#goY@G$UuY{uV%m@hfMQ[ֲ:_XMAXN@Is8O΄i'lζ^T[|g^;}$II{!922131ody.<Y']f$O4׹qNX%)hr^#dHѺC"g7q2zI6fٵ𧙒.ZK/aZ*mvݢIoŪc~ݳ陑%I Hqs9;rמ%!Рk05)c$p ETULZСMLS$e6rhq߆ 4GRܲ@UV+ P)H phɃb+*U=v /_Pn;aOe5Tt&RDvIIZ`=OZW AqEjd8Xx TQd(TtێaY볢4g⠻_n{u1ǙPW*{=Mw&}m>A1ӭ^Í]V;+YVe-b§kRHJ) K9x8J$G]L!"qi-ZZ((~솾ZQE)xIeܨh2hrKP7o{c'#gxEIAHE=p!b$OSDyٗɜJ,82r*:(m+ =ٝݔ!~yȎ}Vz]!IËnَi$W}ЧR+|Z=ؓ < X+F&n5hH NGєF !z(*F1k=(>SqH{j/Aѐ֑V#k5f Txl#E7蠝u>WA-f_8'x'i7(c(hNWGTurы%23MCz*UOy >aAȟi"s0(<(⤡LTi~hRڷ͓:^kc8=xR)"^(IsLؠrOŒOP} FN/0?TpotS.Ƙ= | e8ʹ/~2Kt9\Ox~)ujK|Ew?~-^{9i~I?ۻ^-^wKCFs-:?oEazWʱ KX VtU7\1XQ׮;Vŋ*V-se y/,Q}"N$Kos0 nol"er)YwSRv.2zbV\xBE-Pv$\[ D3?ws 8ol\6gR{X)S4{Q|Fb%!> !s}-u/AJJƝ7y$"Mty;`}Up7w2TkH/)5Ãˬpub}aU6F) +OW9/3Fz~`frfݳxǜcwsU-lIny4wLc݄w6c^S61r!/{$f_XNqoީ[DiP:I?̠vp{hHF&*PZK5BT\Ŕx A7‚lI$_SM5D0|٥1q@I Z VIE s}1˞YjQ+J/sOq~FTŬ7Ys1^5V ]U&Je[6H ߖhi 30O#V3@^X|ӎDK-ǰ-3iVUSS]% et-'%d tC u^ҊQdEΒePSuBU9/PΖ Y tc;k&Άv՟-9ݢuÖ4 ,@Aī(pN  1e)\L EpOIQ҈6d 6(kͫ*E/fMFV3O4'ҬSX } p} ކ13o#v.0qpMVe"4Ӊm|7d6JcTS'x-mXAT/OYm*zk@f-ܸOkҪyI/Z!ORμ-cRΊ:$:ZwMQX! -P(bZI,[D辰b&2"6$(:&9F7̦U:hŒƁj mu25Iɹ(QND,S,ж8wC[%M{aoپ>j~,-7bZ]DS1t}YFwY@ܹcl%Y! rRxY*;pJdiD&jBXb@%jgmn Bm>gE@H͢REwLMLi /aodzv<9X%l#|olpмHR_d ^›_zr|_smؓYD銐6I\}Bwu Y KxZZgwڬټVιxۦڕOnQlny gޮ-]kd\6{pp;.|>vXj{+\Mfi(?~;T~ qsM4_ {gĵx"dj{B|E.◙@ٿPʾu[~/YSv@9+Cu6(t١dDJ5e_㲷?`3W[פ'"jnj"ej~INb & uՕ:!IfjjuJ8a (.f,w2fl$ViD M͗VPll +M-MWNېLOzava@>MID EB0XHԌbBb.@1^#]E tElI =/]O m*,L9Q)mNFDTkL=c;6ӌB>zU<"5+7 @fp~'tp3Y`E]HksbhFEH0!KY]tJjEҪMmld` Y-epDWR4Vd.P̭)AJ;뵡׆k Af)4`Z%tBzH睰F&܂%[!VfQg˾&eʊa>`X0ȶY8wÞ/PTǸ5[[}="68x#ޚI(q= 6 Vy3i`5"Je$t$0 JZbɂ*@̤QZ)Ff VBT!BcL=o3ՁlU[i}mxcQi<@@Ȃ49Id-Zbwh~q[a3?u)5L~?4ےO<+ޏsg%k @~;z{b;w,,hi4SA"d2J֪?LMX!և m+ux:i TSu!Oz ћ/O{70 qWw6K VCQg)E `Ws(9|Uj!wA@ $W"b76)RX]>fkK, Q#nY*%:ykS<jC^uCa Dty)/:";Wm>iO&SJ&a)*kuBR-"{6DZs7+ >{b}y ~ }cd?SoI($, L@F o !@A9/$?;A?ҴQbvQ8!^M4 Kg{r.:8R[CW`͕6PEJZPUTDY%@F Z RQݜ7NXQ.B=b wE}bF?rN<Ƞz=:0['<i.yٗZџzR51Sl>RS/&#Q]ǩNόJb:P2BI!y=oZg|vg'vǖ?ɈuDBs-6 Kfq-MgF.}J3|jq={wrtw9pBԧBqxo1'duc)o뎈9ѥ;Z=^kS2??/.`ȿ?Ã7o3^<|[?^Ma?!֍d`H7tn~ Z~u0Q-'zv>m7GnJ>&׍^s4!:0|7z{2=+G3Bh_7X{-Xp A{ u)Je^߽zm>IEn-aZ+%ީDE}n3r#N?3bL :HRwtB`\Er8d~%wyܬhZ>VEu6?L&ǣ÷?R.;Dlr۬ bT~|&|DXJ3ۯh/^.(I䘓2Ld}*0 A]૝QynfCs CbZ Q!j1D-EEGF= @`l$ r9uIt$ЛrBBYl[U EU1ڎچKcXM6eEDmmRIPt| e uV\Q 3m-DF9*v)P F% Q,*;KKhĹ̲.ym,f+[κ~l-P#fc`ğgמS[TUwԵ(If$֍#]h;"w.9uIe/^BFN,mpd~]oGW{܏ssp v %H(߯zsHjYcgd9Ɪ뮮"ۜO&7˔ĩI,,rRDIqEf1:9٣hxYMOx-_|zNODy´, x9nr\ Q4hOsmu뉝/wǨtC͢SwN[n.%xoTyvS޻u7_p%7t~(_y1S[깼GLgsˠOzc._.9Y8w Yam͕oxVke Z@@B#568P?)ewvzs1/9JKT]hJ&2@1xrJBԎp!^w>1y0;kU&(6ޖp {%Ke8>3ȓ 8QBl$x&J][L( <)RIkkOl\K튚f)4ԇ3ɾyϽY6G \Us|6Q8@+%;xnD=ػᑦH[c3&,!210Фȹdj[jlJ5[Xmdk ee[([neAiw\_Ý(x~[ltsAf2>$"A33!3 j1$I}L!Z=IlJm M&nǬ\9`lm7Zl~:7(f_vѱՆV= ؍pW xLXs4Vಋ>o1 ܡc~LsOSU{U*NZ ![ Z$a&J6F1f"9+jl-_Y./"Q q= y05Rh&e/ j.I<ic,pKL5I,/I"qR p \' 4r# #p‡1"~v ;b/.oңIj hog .0M0/rnN&%d88B>#~;B ;GH#$#4PSY㭵i@a7.*d$N& EHY(eYD_Y l2602,1a&H$ɞj ZLf3su5qvty lU?0ɸn~ R_'cCZu1)!%p-0<:٘|"8# :mP6z3sR7<spЇ3 s$cߦ8ەx/+P70ew.Cɇf5UO}.o,-qݍ vrȗsEֈX!DQy j8 tԧcq>={R'{F֊@ST ~AqGx {]a3ţIqCvђaHKWXyCwd AsS#}"ɘX4.g+n3bCriM}CoNEV> #FehL 6S9zy!dm4S7:}&;wU.%Id@մl,[*8k87 I٢vV >UE(M" gIF皚vr)B9F[MbRd%}|Ju q+eaP`֙ڣ6Ik"XIZcAjL)b,0dp;vva7X2_źpbߓXk`ꀕmv~_ObA~t9)le-(])Y"[2k^x 'PQ9QxUV/~^_lX]iW]2i`lx4<$d~D<*4i[yYp-с,g&svpd\ә?:_<;HY^[o5K4rpc`,yL%2"J+ne2ҤMɹ~O18$}lר^9dKԳWnzlZsFr+BaBT5L*Gn,vIsq91;R,VGey޶T,m/|syqMpv_ DœI>/zV랽@Bme1䓣78YwiƺjYk:FtWsc߭PyGjd(ҋe&NNV=s|68:klsFvVH%ҐRtʭ/d|4]ܿaM[?k{V/fCZy ke ڒW IWFG"x> [~-%Wj%|If/.T:naPd D ");\8 -G-pܢc+})(Q[c9kSp,C9r;e-:sò|kk:ȧUrBeo*U\{&؊4ɸr"8,h/2S֡1͜gh% eUMOUf 5G)$tmD޳4nHPp9)e~gP~{e\ѹȜN d B2.20 1SZ[#w*I&0xeԁq|oڕEQmE+q m 3>#(b B> Eфx/W-\iD95b7N7cnJ\,13'c<)[xt9igpx5׃iyuՒxel8 A5R%0 |8ɀ߂w.jfLN\Noa.C(`zzF׻>azy[tZ8x& *Wʏ4#[+H߽;w=ųd28@ENflv@kvݱ^AXWЅU>)tV9>ȭ ضxn[\5nL-lLHKfՓ!-R4P2ih"5GDU@]*'(!7%BwM&Po9ՠ\u%`-LĒ)H"*epcIz!&霹*kZI笑9Xֺ?{N q2}zP<:,כ[v0P 0.Ð_Oyt64'GI8hbݺBchw%4pw2eQp\'4lD %oA4p>h؀n!NpCIs>X qmc*3C=t~g!6xyq(SJku쫝RD#0"hr6Ibr bsQ΅IU"8^ =wεs1I@ he6rcPAK$ʁae &v ~g}z) Wy,Yj,[S,0ή^&`oN5(NE ,YډI.0g@!B<*=ʧOY#ud<ћ >a!$J;22){tѢN9HPQȭ V"Wh.A>"k^H*zjSe @.r '9Cў;n\L4 vG0fU` co6M3g«FsV.=WY lVcIˌwä㍗ ,@eMWctU6l*U9NIM:Z$E,ƭqnQl> O>vSH~tSucKvq1;ҡ~^.^hg5(] ꌗVo}IPw~{0'gC 3H4FdrVPhR=% >I;A֩pOkQw) D&=J~겫ͬ[Ja"yzȡ"a#(x 3욾?F<fłJދT3U ,RE,T4 n%D E5üxDLJZ@y:XmP&>')ڎhy 8v5׌y (C%zbfu@Yi#m`]kE)%,,B< 1^Kw[%CyRN&a*@h +q BD^VP%#A*"SQFu;4ԛjQ#->gA+9m- qJK{6s spBOIrէ 즌VJh M^ViTX(3 9z{R ~Y=Q֢Π4@=* J%¾^ajYzJ4}gXxZ?:ҹNAhhcdc t cHħ:zi׹s ۡs ۟K( eRK2dA"m*JM.p.&v}(K1FJe II2kǶc@7Tq+۾S9۫CdV Q_ S<\2"?;da5<8\%jhMd U p; XQ'PF{9;,xRfaTPeo EXBVC񊤒% f, ,Z:i3ebm ؜Tr9kD l¨1*ܕ#XhyWyAzێS2@ j Yht"-^-kSJ"%Hld"zxwpCTYQVr/I{RLPk=ΡOC1 L !z VD1f\`b8_\uXYk<ѵbe[klF U$t %k(˦[0f!OHD*j_DK[{< /+l*|Բ#m&H3$Ĩ0i6lY[lQDdd6H'UllzaPƚ8:hC6۲t׍5aX*(q#nL4&WE 7nHpy@BDqʀOPLt&|Xٰ!4Efhq cp5|FǼUo]WY=XBo2pJJ{T*^;S-O7JZ1f7mt JĿeWQ|L#o"|%1!|DaTdKY.kR%U))Z2%D P2A(SҔ $$DҨA/}Vl@STxC-;Z#agcy١\װiҁw G^xPfчӉGS+$lI^i3 9%3c5QF t ¬ft"3s݋x8?"6a Z2si,;޾y"Dfe5xo1JeĘCjΥq'%mo<¸sU` \~RQe&M[ %L$bS6YrV"U5c Vc /BMK%f\>~ txv1owI+{nw6ݗw"F6g<e,,H: It2A bS Fa'Pyk7KjpCE"LAKcrH.(0UYߣW:[i8NC =܍8<0fM>[m&-f[A^i\{ <!H2@-x 2[cK SMBLQ'0/ޥzVBO<,Odְ/I Z)xLZ!lՇXFCGV>+߿;Fn#K}*RGmP4̱ԫPDr. D 1 z}=`f DX+dv] š*C(m8%4D4`eVKf,z=czk*TFxc ʹ6eݢ^yloJV˵e꧹LM=M] rojSWrQKmJצTkS?H- >\U \Ur:ub*5p @OU%غ+&hWZaWJi{zpeOZ,^KmpI2cf,9}{4f;]0߀7/磋iij~}!!P UV a}cܴJ xr~08r鉵Qi`QTa7^zYk .dc菅/6kBe'j9 3Fy :8d;fy%2s3e-~@hRޙTz*KRkje_Vv[,'( mRntp *H"ZjXY4Hͫ7j^%ܾyZg1W󞡚'R\U*WSK\))5U%XJSG<ڧ$y]$YWUur,vpǕ- ǻ"nt|s;:@o=/u[9*7[ ^`f8q* 7M W}숙@OK_lVnf O\cObȻMm .uz)`.I|{s[38X x{#O'h_>7`|50D Qdode[;YB=^g~gio]}|c)G:ͨ`vţ|>N@\%P_#/_z\]k .akrf9ܐv:6=ywvB.ALUYJcMN6Dӵ:]RwN\flG6UVzi\BU4F\ &^*5Rӡ+Rm~qdoV{]QXN>aeV+°:jB-ߐmʡE&F ab"TDSkTiu0?-'t_wn@wyz}U `lvK`VL2S5qlhcrFh‚Nv88e+nCvj)+n#ZVZ3܌Vx062,wX`t $WKZS)ȌBȊ5 |*(׮a &Dv>g!%$'N`][oRM%9XJގN9k 4*%Og䘭QO6;:xxdew+=pŲ\_r;Й`Rs'J&->'k o%^ cƶ=O >gyFXLDftBRV%!%"{̆la=V' Ō;9zwROu'#*Abqp-+18ȒG',ouX"v;i#%ufW ZC2 R]A:!tG|i#fN%$EA24ѳ\P(wDj&~  _SGqҎ%0m& [guLۦۧlT',XvxtZ|-J8`g]]Y^fşVR&V>L۸i<(&gjcf=2roa|7ӫqn8 Ū@l ..KMB°lEqW_M!a9MP%Kĩ_\p8<ó?߿W7ǯ{o߼Ot_)vuEq;˝P|6q555ͷZ"3yc^#k^yz}^%ت$q kPlk\_d'o)!q;IS62XRyaLŽѸ  #tJEibLR.M+=tKax27!9) 9C?ڐ#)13JF;xdY:UtB58Ɏka=olwypC<.Н ) U64J <8rcbTkP*EsrCҾA-l#砰C&S$ϝ~bφK>@d 2ș #QPa׊Ľs52wcn{tXpE+5<<\#JEB-W!*7y< 8=ofjXl.G@ra)6"kK(H0Jݮl١ޡ.Q%90*Î4&0X.X#dQRZ R).T.HRdY)_@uT,p-",7_d2N0F,i抆R;'.Cv֔--E'}"=&^ǟ"~$rSb.{,4Sɽ$LƦk=eG% hKxkuw> j`0D2hi $4e /MJ.%@M)zHq[,Fk2']Jg ]^V#gXnx>hc&D `NfΕ{ E*#Nikx5Qi49=Che))`@eb&bbژ38Z:LJfF횱V qơPVօׅ <͙dA,,sdn^ &G+ FIpdʅ Ū̄'Bb Ġ$I}L!@P= )ڔR 6LCؒ`vb-rn~2&fWv58TkCe N' 8, *lчs. : LsOFKUaVm=qTPLtML,hM)Ƭ" $謬s>lUâǡQUֈ׈FIȃ0q8/r7) xQ$ 6gmbAHB[mֆh@gC0}x*bxx U9 )Q+l^`D뒦 Bc4A|C@S4,o`ǻ/U>tB` ]TQ 鍲P "eT[-R$ cYJ6W|RRAKy͌8l\03Z'g)N?R›7<1}Z, M/%kѬ RZ94@Uޑ\M=޹Tx\uPuDA;\^G xS$BfI&.#3)G锃dncBty>r,!duWUtGp ,͵}٣B/w)uǐ6Z{ûc)0]?osC67tq'w;Rb+G«ȭ7b leaHaR,@&(d=l* VKQͶn68 -vXË,J7SBs6|xjUy}K?u9 ׀a)K>&e4V +Dr(4\2[7_. !]tV3uݍQy*#nLK >79D+1*pDit٧@ c >Yocjx)<'*`ETvqGާLuO}-ۭǼgۍê$ِg9Pc8cVaT2GT/ ZĶ4l޹:_ԛQshoG9J@ 1C`,hg_*=v`^w1)XcwS]]\km#y4D06'#S 4"R<7q;)/rUQ8K1S[<:+1=fγTcĈ$Dv7j|n s[Xd8^FۤuuSDZ,x Ie{aa} _.-8r&QTcV'煈Aq20!Li#*I6zWLM?L]TA~VAwi3.L{$7{ x.趇 q4o>E+j@kguc]̍`p] Lut@R.HoURА5SﺞBL9I |DlbܑA6jQhlglW0&^z(:ާhR8xƃ9Wʏ#i[-H߽o{x ({GleOFʵ@n_mqn>^ɉ|#!--5*_!?.yn~Ia2 Kmr@³g唗~O޸LK1~L]6L{' ćKip}ỻQRwld9EKf(+6H_rWYΗjdz |/KpB>(D['_pIB^d]+t^vq:vnꆥ~nl/G~n ݮ&Xݨ.}g~>M0xe+}nFʔx7#>XJnjo?dT8e)!q D5)yۢf@kH=*"^_囧E>9.hNp4<*jOzEڛczaWGuRqXtܿ 8<e,fdrHxGNcmG$d6ӥ ZGK!2f4j|󴅧FXp45l5q+L^joMdHmQ{`2CFa[Dg5~aƘ@JE9+mX-Qm}e%5\\r@YsAdQ$#NYkga^EU^Ђafm^reTfo[*/LN ľX\{HX{Yq۹e}wثU#t51NVej[XF-nK8lQer΍"*49:(#Gya[NmM Xhm 4wFm6*GT KIwPӷ+7Ff ^I\=9`X骹>Euz,gɵVo"v2oHL"S2!H"%bV !1Bδ;Wҧ.:n=ع`S2(hO!)DLfL#(嵷Hkxm6_Z鄒1 &*P j4!MZ,3BR.uMC =66`_(_6ʓ%x>Ȅ\L{S*P 2miZ儑 QVrzJG#DxZ[72- N4e rl2 0aa3P(#DHD"# FT8GR/ A*!lhڰ5FΖrG fQ1*>r8)X9":e.)(D#g Ra^2thEI< _+-NHKD.b7uxg0@gT:?ZnkOz yFi\ ((ZgMLJoS\2HVq48a2JxR%oFۘiA0Vi817J[JR}TX`!#11'->^e&(C+mͭmQ%LP1Ud7ˊx0RVo Ϲ3pДu){s,&h-X+  ]0lP<`tC8dWP KAdXإ!W*"e.REgn $R, 6B&_9RǓN,r2b)Ke$"_jY@]4aǁՊB]0jiS7NgmkQcX5T_R)2)'+? @9K*Y0R`D$| JJ-1KCwpxT`lJDPRL8ZBPkJ0Rklt Kp3"اdL/%%\2NXYht4X9@'RR:-v.5-OL .d̕8oݟ{*zv2B͉`SZNNrkTu+oNIݛ>&6aȆV4ms+ |7_ƿ=dn/vYf̟hr|׀?#( ;_ PUʽScPǃɻnϞ|U U/4̈́0<'KլO 'N=?}~kopOb8ӝbɇt&x;|׳"Űf_>_ws}b_0vDNz`d^G?a**tg åG3j\`NNIѳ)`˟_>g| —*Vs_Z S|9c1;&BlFctor|$_!'YvUSvߗ >Z XF)Sy8 pgo<>reUƇhe_rg:E =OgH]\GVzQ$/ XMk,}JCĜz'?59=-^Maj~Q$u5TvnNր-sQ {`4ЭjziF L u-C?}NC,_:duR Ɵ}oz_ =q?>vSY SHg<G X&?@F;I:8;XhL"z;{F [$!.`_s%UЉ @˯¤`ix<%b7qͯewxVDQOir9*U~tRsu !>o(t~{~mHUgiO @RI'K+5E^X Mu<[z6HakW3tzâֲ'ղ|\CZڼ kYR*z݃=}kVe)iJݿe`=3~gee#v E2R͵CxCW쯧3DP]p$Jf P]j3{C4~CePo'uI}+6Q\EZ44[Uܙw"XZND4*:|:yr6#EJ.7D0BJnXk!Fz'2 ʹA#v/ڵ־sWB7""R2 ?^:-@Xh.?MqXCƷJ8*h7v=|л_t0.2vd982{cŲń9rԕM]q|Tc Q7ar:'IU-<$t1W_WyxWŏ9(X76 PEP2~r\<Ȓ 7t^|E&浔Ag2Cu Q,ߨs6!)R1_ǯ;0x_ێ2` V |>Za4&Kv25^]),:e%B>tMGĦ%ʕWwAP}@^4Lz0^t"E'j=t/:QiE^tC=mhK(j:R h62aLkڸLA[˶iVlw-)$1Z|p* )2 ҙcw=׿̛msю53u:7B04CJy]̊nD̷cveޔ\!a?XϤƶ\kYic[{o\nunp 4wW w5 +4H÷M UUvnHvBNTk)9juGfR_*;CۘcV}z{96 uQ{Xnܕ>zVjIy/7,:xV$r8ab~|UN$|/NX;9ak`c.7@QOLW5TѐPvfBaBGt5_(6?{F`!sc`^2IvyIf08WK0-jII1:͋H(i6xlIjVwԩz^*$8&x5*T LUJE;\`rh'X`'㮪t2,Z8j>Yw+A\g_V6~fwx9C}u@#&YOs/g;AW(Zۦo>Z5I{j_0ǒFgTjj *nv}iuD^?nn(g~UΆ_U'/4m[knox{mJL~ލ~qEK ۅޖSz Ǯ}eZu2)g^i@zThO]U=wUŕTUVپ*z/AyBJ('䮪dO]B+RlqWR;s2ŕ(N]UiEUR]+s[$ǼٵLZVgusMJio.(v 9L0H-iϡ;n#ݙ8ϷjǣGg0*e;Mf[,/;f`6g;Uk 1Y-Ue7*~]T)vQ]*c^2%!=U ՚$STFr:a=P؊!m@c P2"4(pC>k Hg.l[ 8"ggCW?{7}jj ⨲xmz֓c*ض{,|1[6gmNNl[w>״Tt5z9O/_V׿r5/`-k!жX1R: ҒX{0OW&W -NJu򀫛[:3$(GcøJ1c,K*eRPל8)ht,iZ``V Û!cRO/?6y?ЩonVN3wF}-m_ ?yy:ETJH nӨbCR&dEp?s]}o/YO-"doS"R^ 8ez*btU0)%U\$RH3޴N*a {S#XIf&y_uڙ8;=Ld?zSz8p>\ͅx|6|n ۛ9^.y)*1$[ؒZ{L!$g`6dbCDNn Zg|]qG4yY*Oɻ2BQPBG#=U ߕ8[F+ʑN־>%BJ0ph>+fy4fi:n,hU*t[bRcEe<"f6tyJon4"% US6 uCƉR(|)!uyK.zRxi; &(m"B ll%&(ˮxEL:hB]Df`R'BR^bgqZ@NFrg<ϗ~O Yǐo^/x\izc|JD4-ff bz}$zmnoG|x2N I lpďFM7E(m5٨iX=lPuzHuUJكÖCs9^y;]}VW\,m}a*G_nq(گm>P cH(IC%} 9i%IRzٓ:\ϙv.B=b؂ٵMF@aDaȭ`p H5z)q`e7Pѥ1FHHz)nF_:3FʬuL P6ri@`\Ia{!z^P^O=q2=~POtOm)|0: J(ujFZ1Kb,?1 J}j( ]fqӕ_.ۙ"}e^}7z*h7k͏˅ɞLg 3Ґ.YA&h| RILaUt޾cHC2G> ۾0TwCrIFthC Eͪﵗ* NiW!l;[q|N2;<2>q[$~LjdDšpS _{}6wַ?I}7 d-=A~[)ɫ;UW\^⛊ʇX*1HCVR1 E-\,AeS% pЍ9]|$Q3Y'} |w3KF(%?IqkzWTzyy#*HAfl2zXb`|f>[Sbˮ3Z.l+k`v#PTɒ!JM)@Dx= <7t4>K|?"տ_y^?\_{R-dZ P%N"h])Lku{X;Mgn;/"|HX ?Vך(k||nSָ8Ƃh{JבTuku寶8~55O[1 Xmv6˫5Qq2xs3] SG{Ϭe]10(,Bk.kk$jեڛ[jjCߕv`2[ky[E!Ň"ά/1[mioid|:cnNރ$DmvkυzNx#M$:k8J #J6APbQkNPe,ig4YګϏSZr<"R2 tZpV\Ljhqk>d^.;:ܸUM^3梽[9Yg wv6CnoC)9Upp~_j݊h^~++Gɰ%#ʕ6D^,+r?_wƔN[EeFLVZF.FHFZı$p3*AqZmޠ:h|0!S&aX5(K)µzqy;(|.qt[.e8_lI0Wܖ!|R-cʐAH 3BL0+!aYBl>Gcl\]|,1H(L*"Q$YTqXŠ$<fp'XO(-}YX"b8z0JXsEqϹEP:nFmeYf6x?eW牡5mQ9#9搦F  UXA6ҊIhBʻnrG/E;bcx`H(@T@yoCEN &,|u)".ԟ7ߗ?@)0F͉\`C"I$̊XLC XR|ӟeJr&`Z|K, ]I}akF;m?8ZVή #p ȧ('R}1K$0F strtjFsЪ[7tY~*$Vh\c`ŴGn>gٺSzeﵻUVLґ{VHMxR +$bܜ_oT"ZRؾKLϻ#Yp:@i+Jx }1Oj"5IEK0S_785ŅLNa뷟^?oۋ7߿cG_};\?G?oߵ@bt-s^u;[+iK^Wł\L3wkIS2_k#mLN"J! 2\{EQ-іFAN/(PG =>Xfqo'=C.ezޥC.Xh=X,s)i ܀Mi8ʅ  EA"?\gFgy6Ow|+;7U{#Qw(l NzOo5aų Ƴ=oz³g 3'Uy%|T"9.; Lw_~KnbˏMZ"W.2ݤKt0uQK&R/ P^g&xIo?5Y=wKAr_@ALPv m^⬮lGP /ݙ8]$XͱƆ1yq]eQ獧{x-e8?Օ3OGU=8zc8͛gE}:uQ GÃfcqBԔOmQ }TPyoY);{%O !ڕrdX,Qd eJ+gI1٧q٧|ܴ6JL%;0`dT:陶c1FQ\‚#ۏ,qqg 䈑*>p减7KࡽCY|vin7ɘT];갾!} ˡd,% dܣRaJKnD$!JYܫ@_h&;svȿ/vB]Z>pRzrZ+ I%!z8f$vHS(JqTQܦ]IGEdQ0A[R.k%#^٬Bz,97k4êLziu "~aZ7 ucb>C(>lνD(dRJ-^zi^ܫ{rG˔C3MPjz-XI<*F.4JR-'( b#C0!B*Ca k5f,`ZF MFSY&Ζ\?r4לl&n<9#SuwmETK[4!xe&Xvt ox2lr8v^rhMj$]91I-+1I[ jMA]3r48ͮk-_Tظky0״וn-x#_qPy} IMR ~b k+ qfaHB9UaaAP)r}ټPɴsZWϝv+ 'H PBn#k3086dg w͏""ό|{D\$ĎHq4pnTVkLyDրц@R U: hINƁ3uS/ g,@XҠK0IrG&fD|V[Ūc.YgYr[\qQqqfoS )urH# be}0TZMTH0<)c"ҙ5llIZt|X>0 eA:"Q˛Սy*汄Uup1,D kQ eXP0mBy6Z*% ~i"cc2e3cPp3hx5SDc>$a 5b3h6.w|ll^ M#jubNcIU˕6ޭp+QՓ^__" R-{lW8z0ETf@ʆ#~Z(RdKJ2VZ?T,kбŲ6{AZɵ0ԥy\ Ղ"y ZJP)N %vhY^E]e}+/i텗q0Do`O7uGHo#>:ϓI@e$ Jr B} $'WI` \WIZwj)@V]%IJ`,t䡁NXvw˦(xDZ]ۙN%Yv^[2#ەCwë{qW :J\!w^{1XɇkP.:pF:QĚع*7~~ruq3?7Q׻T ǖ{@N+r]]xʠf09X8s#ݜ͸f>%]\8M! 3\ו/i|}~q}s754)A7'n *ϱe΄V-_?}'?^^~^,. gULrCSScLr nbiIJaR_9L->k>{9!6e@~S 2USi25NflE=#  [GU*l'7lyp-d;6o-ݲya 6CO;U2 \; ZovW]W{+UU0ҞQp*W+ +Y+WZFZ+:pb rW]y\u"\!(@`o0rybp}>JO\!qz \Ap]9aE4 گ}qUÕ}ĕorWk\uGzWJC\I a \A4 \fnqUw*8)]PvJzJf c.Y}|ݻ.؏.7Q0 F]WiE7iVzVơwAg6VW'bsluɽ>zN=ۻ1pNu9ѮNx\dc[{)gɦ0mmtU5Va|&rCrJժlToyEt5h"%jaZOvVq-ݪU\O-,@ҹݺVq~*>e5R~ \Ac= W,+MLD+4p2䃕peHԶɻ\FUW+nqW{+JW]apz= TC\YJ+&U(j:J\!Fpq!ר0 ZkW]%֮W{\!3D]5_}{qރ@ 曂A_j~wsqup&f|~?8 Cv qJ.Wd{:^M H`0\Yjuu\up  f\ubFԲRW+Amd;7#4ѫn#ezdXwFvn۟Z킾k}1/?i=pIe֡+[ z]'\vKOIYOU;vOao`o:lW]aph\uvW]q\!4`@8JWģફTkd9jqe8`Qj&7*ݡG\dG*!X+; mSZ6>Z z\YkW]-ɮ㪫4wr7U-"⪫t=ĕwʒW]g\j*?jq%ΰpx\uڌvK{r:WUpcKDN>O  8gH .w]O| Eiȡ׳ݩ4~'vi'nV0ĮoVZ>=5dk>ۺU\K۾^\~=aK kcwVQSU皬'Wx\A-F=W]%W{+Uʖ|\XPE S0{?[V iQ>o>ϟ-wͷ)[Oď3[4>#?.uNKEj v&jNRQ4mUrbJTt";RdegH.S4?~湾767+z|U˛xӅϿ[|ηù˭\NHY@͕UZ^ŋq\5L6hTlz$TJHЅ1*՟o1s1XW`HVu6YmHK ͸\ TA@1ׁBlԱD>Z2W ۤZ_kEI I%c01WhF ibK1lr>iR-VL _ػsXUԚ$moV)E7[!IpRzǪT seGCƦ#D3y(R.bU,i=KF Ȏ +oTd!o=GL+ϰV+]`rb',a[pQQw ^ T9kUsvb5Q& (GNX<Īuɕޕ82hA'C1cm̭Lp98Aw\Zl 0ųnJUpTb6ٰ*p:#d4dkkAL-WA&6bs(SJ bjg(cdC2"4pNp$SؤX <!TP WEIpS`Ʋ]gdX2)fP`?FnU{C%5a0p(SP [&6)Hh ".&7lVHn:nMI{tXY7MhTtOS*?%}̕}Ρ'8TR 3*QkVFfdB##]/Vj!HFLV)8J8I9nXT5YVRl eJPzm쬺r {%+eATu]0%PSi f^g2,$G !5P^i-%ћa ev1`{ҰZ.jE}I^KȠmx U_[]bB\B_1k <_#9TrvND(l}$jyeg|<˚xyoď4JƬִ<@mP 3 f;wQVJWmKl:@G*N.GWDҫ{ wSPCs ~XQ|FŢN A-)JH)DN+̫3foXZ0=Ӆb{0zOE"$H>zt_U(#p&GͰ 8_}I,:Q' ;UϳwVU@VT%gF]V3#Hb W!t8;ϋ?wN}}6FF|wI>`Wh"BPw|H0ݒiE)(3JWExhM6cDE0Ta< LEX!6vWn$؍zF,F,iZ-E̓_p J(YCnjBfD&g7$Ce@XPqY]Pբ bQyh٨ ITT(*!BPʉl|j%[W`~A,ҍ,vOGFЀ,nY6@)lVTR@M[)20F KX  y+8C=Kp0/ R mtiXq0=zy_\@t>_@x6]>8K^D۷@+Q]`lDGG<Ɩ8Bؗה]nI˜qJ/ =6jXKiRMR^{hKE23[Jdf䉈'΍ eizj 4kH繨)@r8LY5Xh9@nb@bwoh*3| gp r7xSArHfsȇ醇Z+5$VtP,B0e TP T"' Jz kdc4B ?z_++&U.1x1YJ4h`I;wt,b:}7XVl2Y6}_obe /tX\·\3?|vs,/pz9ϷPfyc'pvf418 zD|Աd1h0޶ Zhr^q;'=U!J =HrŘ u'2脜@Rᅱ@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN s}r|j!@W86wJ tN `z"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN rt@0ëG\ hE@RZr 9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@ŪY@8ߘ y'$'i:5Z@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 t:N[F[vbe{sjC/ީ<_dLFx?LǸӽ1.a^#:o\RqKt#R4}+Dԓ QZAtut=Y]7eo ި+ ]+D< ҕk#23zCW)w)sFtut?`)pt+kY_ zuQ2xtR;#L3{CWWf uB[ DW'DWFIyu-Rkxx~  f_nܶ7m5W7fE#Ԁ顖C&W>̦Zs :G7\421dJJ_M=-Rn?I~p?.MV#~(EhZAӒh]~M_-F㖗Yy87L ϫI>Zniq 8N P_]z58_o z_Q79uMd8M K7O$_pZY7j/*~xyH9_g mBѻ9[=o,=hv1\֪亢yL 颤w 'OًRRl>&6M )elT!pjNB)0\c{$72 heeTd <кGt%++y_ juB[5N$>Z7th%:]!ʭODWCWJ'ЃZ}+%4t(':AJ+xo֮S{OWR-S+#E=+ \sl~hDW'HWVjtg?ͫսYB QZ3xtis/G]!\ٛ+D;QZ:E{DW2BwhUADi#IҕтNNfnP(8)'tA{[g }>zW04f+Q͟*f=}JmN^/,fG~Q/eT?Nf׺EUmTb: |.^)zi禓}WhdV&9'._QRJv@"XcguG^ =%C);vD!Iv=W;#1zCW>QthP*Itut:w{)r{zڃjP[]NrY ?Zˌvyv.|M9.!˼ ce֨D_nJ t.ͳyfEka(luP b\8'!Ylba~GXT/K.?@#6.Am}v˃J`sիؿCv^L0] w~ZWu޶5B7Ȧ?j^h׫UķMr\~{ l7ݔff:cy=f ü4|8:c ~_ A܏dVZ⦪jI#ιTELV5=˲}߲ٺ GeXL~֑뫶^M&0]:wk}_v^ŦU"r5pn)D7d|l>PokY4"wCfК`.HiI7M5pfUaP/omt;&@-NPR JVSc 1J%^LRpE0m@rdZ|>ۀ&²gW,e]p: X,.@Q\3?|譒P_PMp6ǂvilT?o 4 .8 j{bDa[bʽ9 gRb^q;-GiOp8̦cfUϛ2cePҚxcjN*jDQ2hHx9ƲE'WfY6T}!]L@BXXv+cn I}:ZmRɩx t*dR)Q9F1Sdu.0ᣎtU,1+-=` 2g5=e~Yn8(vU]c% EAi`HAāfFE ^+DIITBmHcvB[C)\kJFaL"4[k,X!Ahtg 2 Ad 8YAsxNFÂ2Aeq(I$b.A] uì#Mm]LՐ8%d9ĉEHpP(9Ta0G; W5쉏\S3布ܺ  Ы֊ `['d>W&0(兕@cclVCuUcmAjcgܸ3)gZ/ڮsVe_^~Nr`4A^{%;w-\ c1=zAu/ hrSMKL#<iv|vmP,>I,J8( ɕMI`9'\I^\eR%]S |iipK6%ء %̌Ft a@p?`JB*ݏ'rQԃ.R)L)0K:t]J\.;X99}-MtsEVW׼5ꃪ2P]Ny],E 8`yBk ?TP܉)o" UA!S67ic{] &HD} E)xLFޏ&a}܏ 4-"w~q;j릘O'[qYw~Z#-Yanvh}5^/ q*^j=_c\Y}shq1*HfukZG=̕q]Ux ABS1;|a<@U/Ea.˜Td#{i=SWR8WK{ ج~۹.ʖ.MƹX)yL 8ʏ0# 볡[swuo^OM';br;nz֓z*vWb}y=`o&d۲߾cؚh[{%u,`Î#a?HObzwJ3dCO$>/$JoVTa(FY]IR4FH }Vkgh齳fJ ތ +Ĭo9Oc~}`LdfD1Sfz[}H0R~Z7xrRʰdkknFej_6٪4Cd7[݊CR5nɐ9~áD](ТM$`w[2P[P9a Xome2v  Qt.|B1`=oP mv.yKNK.:e}66 NABZKL1!ٛƁ:6ՂKh!q"f:}tFΆ(PO?/v(珇dν?["nOlM6^ݵB; ƳɢM}L#u;K0AD-&g:(Ζ[gP7ǙH\9P¶8!mr*E}> lNȍB HBîwtE| tkZsO8g<%g6`6^7MY.$SZB5Yډ\`ڥޅܒ )w܅s!r'\HfcG#0D% I=hQn PQȭ Vɺ\rs#g P>|vS!wlvFhw䔏39秓a2k~1Û8I?NXߌn,IV^}&{[ͻZQI'Բ혭Bn }n]W%Ru#$]\̖l.χiM&eخeҚѮ_9)*N nF%_ 79abBF'B. Cg0c9% >[pvqھϴ"!Ƙv.waɅs6XvQ.E=hL- Zj-jUFk URes)+bws)ɥ@&E\^gƘ^+aK7 .z\<8noHS:{m@r>YsBx%DmKFetFvUѡuߊA&x +j2g *lRaRKM9 %]$9)/{~F,N۝7OF{9:i"cd QfrǙWsmȺf`qS9CAV2'FN0p&c(76%c :X:#gC9k' f cTz+)RW@fpNpXm;g d%I2OLh$g."Х+oz A)Y4 mD2.(r L$Rȡ a'~ IU̍+q>/74sNag LG_|51!fI8L^qt8fi~6ٕĝĮe'~ 6,٪,'WhQ rݼ퉅pǂL̳Iu!Φ~EvE:c X6BID(A>$ifRh.Ƞqgt0P<.:'"qcv,s pWR2]눚Mmrh @Yz3Mh{§RYϧaQk@D|;|alIL&Lk_wLٵe2!֚3aUr*SLӣȏI'l G5NȐ.ICJBt 8dR.df: q#34)c_&mDc@(#P`fdxdDFIS:#`iC"#;nsj7LR6>5Fƥr xۍO?xN}T(Aʔc1R E9icTLrv")<췳ik Mʠ#wA?{ӹþgo YVA+[bʠr~)eЈZŮA+T ՗A ˠqr yשisWܫ-e~޽ R!HVȡ2 ڕ諝}0Υ1eO8ZOi8ۮ/i ٠MDqY6Fk5Ic'KQXNK ن챯ͦK9\<| ~yF8/욼45EV`}$̊o/Q C"9 ښUNrdQhNDI2x 5V4B]tW b" |0d_h[`"d:d/ABɲ>I5ϚzZ*׌gdɇ .JKwkA-=>XC=>IƋU<,0/.]΋dF\M$uT:.>5sL^txIUZkOMI(&_!dB6xN+˅!2Cd ͳ6Z޳(;"I^:򮚝5D}(! ![Bq>O/בů&gQ:ڲi\ZZ[4xJ?u}IO/+x2)ch80Ƃ~_0ɾ{{SA[(a-qX%- fKuFuY?sm?1}CۃS\@fqv׿?O FB)WoiȃռI/ﯼ,]UQ>>_lVxkoKX_/ڴzζg \NN@89n.(\O}s+fU;RQyKLr6 J=ZZg'?ǓoC Z(9*.ilI]hS[/,fafdl9]uVWZ?;5zt|Uۆ[&6j kfK0Q=a`;k(/h6]|0tY "}vV.J߿\e0,\Ji(s0i)g 1L0Eu nG6V 1dHПhJ3+R^M 0$T+N^_o<*I0^eCpuf!6 G䀸#pߧ~xg ^J,eۮU+4]KaCSVu) %@/W}[i bwU?Ϣ\DW x rԜ C].LER[ Cw[Bʒg8ą`%X-ZI(ghs8cQv]â+r6wMw lޒ۽kQ6*t ҢȊx#qKV$&LjC2%Y&$)9ڊi^P.~ydyGX $)wi02\II5$䐨"F]Ý Ca(bA}Q &P)>*ƥސ88sR AwΣI{/ ggaEA(qpJIX9@-&@N[2\g5Zg! 2zAVU!,htoNF5Cu,AkTd5SJ= Lʬ A 6W 7_Wb LD/^0. ˆqW .tIEn?TqMȵDP 'M%[yE0yXݨQ"?OqK@ 0 kop6K~x7OClF+W%x5^owjcK}t Pо~=f^,OGXp A_ie+*L?wGyui/;}/Jͣu$Hd~z @͇֬i`&C.g.ζ7n_W(Ç*h 8fg޾À_s2/es[_t-WzWXG4~۾CdT,74 Kdl]ة FWL.^i( 4Pjr& ~ubIl1(76]3uh 1w rW +P&nu TLzkʖ&n |D)m\~n71n /yn+kzvVXLooߝnk"[igM*7!lMxIw$f iv>z'#&](boowz{W9YxÌ[mr@jK]m3# /dgSI:ehyxcv- ^Q;o,9wOꢏb4_|4Mڳl8޲gة#2Cp]e ڎ,F&m#b֞S(rKוUo ''J+,4h@s٨5B, 882 GefaW^z' 7oseohӓ˫1x`[mlahm S4ʧL˖ OCF yʖgq!-AeX#ش'ƂΛ¾P 1SWacɹs1'S?`PjԆڋ{"\L"ZY=yB ]Rc`jUN<* z 5BE'=*9 %ǦSDeƂJ<73ɹS$` G%"LD\HBg|l bυ@ksFCyH#'@؈U<o [%яWr/y.ԹYn<;}}~1B dHt&Y x(MkJi2*.Oi`ځL.& Ha+ZL& Al[: 6;CΤ\v*pavRX r,l 3LdN(ƢȖhryԜ_?Gc͉verfM޿}>'9 6hA;ˮ"ժer[Z@] I2<߲V6/֪ &̎=Aw;yg6QUOY˻.X1~w%6zu?lLXq`v`3E'=ےt@<C^ȯ%^]Kq/G~x8i'w'ßvo%IG5Wx/w~ w_ݻ"~a۔V^E'O}HIbţq~Z*{?Hy1/Gஷd W"ثjp%rNW6:D%W3{"\`|5jJOWR Bɺb὎+Jz;u\ʰ8sĕ׮"\`jp%rWq%*b]W>8|ֵ;JZW&.*\W9W,8z+k\Zu%*,!FE_G[t8h;mw{}Ʀ' X)@Gŕzp%r\Z+QI38G\'"\&i5E?u\ʩ%\p,BR6d]`,\Zp%j+QjGA*•\C,j52(*Yp5C\yB W,9+ <NMWW3'+•FU Df]YeX W \ kşyad Fؚ"[̂`Z䂫Ӣ1-*]dφ[b=AWu(s MοlxEA=h5zRPF~0Z9yz4;L%Nl0 0b=umEOr9LZp%jC:X%jXp5C\Y"\`wj+Qkq%*\W p%hvP D SIu5G\yBEXu%r]5Π%;u\ʰjbʑW,] D0y\e}Zsjp%rΣp%j:D%W3ikBMbc]\Ղy@JR5֕ 4u\Jh q_uU3(rMW3S>Wro/݂YgW"փ+ \Z SǕta qmU9"l5\CBr/ y|pEs 뉻z5@Q,W@CM֕zp%rW(?u\J +,LF+u`ʒq'g\|Yc eI%M SprwMK+ g4 2{1U}o¼]чJWB#SI sXtO&uҟ93$;.֧[\^}w/_z}1k)|t?6Rvw/z/??m?Fv/%\?a͇1OڌjӶ!ȅǒwum$,2֗ AyX!yyI<3V'sIYnI+y-fTխsӥ?i^#W80Ileuڠ ]> >UJKeܡה¢ \EW o6j}tEN3<~b/*0.֫tt(u:A"n1t饟rZkBvӡ+rx,8bv1 uGOW ]"]9TXҵ+z9 { ZKNWRIҕƚ% 8,G]mƥMG(w: E +~6YΝAKQWօc+F)Ua菟J/Ybt tBWHW)`;ZW*X(@qG ),8ef1 8--FwCCoDj1;~Ԯ-W}6Dͮa!k\4Nfܯ 60+Ɨ@GVћܡa_]m7Ջ-ܴZQ'tuteB#xa-K+F{jK +{2;Q/ف(f;_/.`>oWo~7otsr!M{}Ps{GWo{_/'c0;d;?#W5or?ܽA;i{|5~~m!֐"տm?+|q>91io60_LS|m}ܗiﻝpwF-fqvY_Fjl{-a|%޽c/^-iayD}*j {w3ebg5x3]}fGg ϑPO!Eˈ u뿼׿+l~nfYI=#lqqtmfou^WY%)ws?nn!,Fgh m0Ohz}~Rfd]JzQpSP%egM z*8]2$%QfoLɅtVTF\U)6~L_ȥ4S5i|3ShЁ5 }P cnN-K w䡊E;ꭢ͙!f-?"TH~N;"DKԩQU FD]40MQr9C1~-?d@W}1QXSZ@ Y;[\\הR#%IE֟KAha2c1^#1>olfVHSzL-x~ <L὾6dڦw)JӁw ml९ʒ&dK*~Z8fSc,c1mFM#dLhwUa}pxĨtb e uqI%z&i7Ysզ:uY(/`TJw!IBާ$ͳzM59ՑJj MQ Ar#钍5 u}gtsaM nN';Yn)Ǟ\1c##$63f Ȩ*OM(- VRjH) AF5K!RXuݷn\ D@K$X'5/V[ZQW!5*dlzaCKU+ܜ-epd0Wz,Ij_*}AI8(l,@'@Hu{PH&C@U+cG=D &O ԙ z/nbA\_1,4WU)*K;L'Q_r0g\n:gI%vq9%h)>zȰ‚SZ ]1\BW6c+Fgyt; `ˡ+kRo(:AJ|~:u{:_ 5/aA̞Wo] Qs̠^)w왲Wqu}ЛktԙW E޻[> 8-b`Fk4R?4m8ՇŖvW7_1^?{Ƒ a{@FC@s 6X g3E*Ҷ|cq L꯫{@tO8kh E}qj@N4CYD5,}Z-C}w>hNHca- ;4uh)*0 Y@oz|Ez2eO6?| inޚ --t\c+,•X *vBHW(~L js,thm+@ɖ:z>t%QbWX?AjFStp9B+DЕDJrDt9BhbWVqvBHWr*6 "'^+@ˈm;]!J9ϑ K]!`s<h ѲgHWjCGCWW ]Z~B]=OҌ/1l+2!L3N(YW[[`Mʇg^i 5TԺRAH[Ǐq4lus R!o<] Ю)2՛T>.eN_Q$*r!QXy~Ov~UsmԒ&VNbaE09,l'87j\~:i7eH,:2tdGCWWc+DkX QZ3+ƞ} rG|fJ;c[ @jv Uz_nV;Ղ%cKOh#$X HH"D<#/PˋuÈ%/dWY%#0 k (%X=?bd`M΃ɨw&* r'_ 0]p8WZF޻ϋ*T*xoud.)'@'i`ٽFVSI%b:mpxk~0@=JIw>r1YYc]Jf=L^Egg˫KSb 9 ʓM!p.qiM}Iհqm sL!iƔF[-{>Va;?T7^N' ``RZ}~Q,7= E8E"pE[AȪWw/na&!nET׊*M۽пA_W}_o~o?Pf>W¿s* H{UϫZUs#Z7YoQM6yE߭*vf2Vbx쥔žc| I}>m[b'?}E?1$n`6 s< v4&j3`jVRFcs'=yaN.E5z~oSlbTN( tB9LЛHU]jy|g&5uT#Ԧl 'YtS:) )l8!YLuFu4^U>Ck1k [F`_WMsS:298}dQ= V}#R"z?O tJ p2}B/=ZS7 M'RRɠ\z9EE{|rתGlt N'UK^t?Uq>N_$HA#uGa˃oS .?iwC+tn(e˶)2eG/]:UL2GN^($.PHy> rd"SAdqW0LS:?@_%wOE99pls[m`pr9xd)CQYhý[P糦\ hFgxQXO#% H|f#MV$!5%PXbf ~ ڨ<QJO58U߰[n]BjMQ5{0F+'VQj6XeSy6SW:@&VF6mf+0MlUTb FTBau9j0/μ˖"-#jo1*'5M›6G.k cNɥdfoM)zA9@q[*D,Jga18[Jg0,u3"o/|E]]7/&LQ9jK} 5$]vcc%4Al^jRɄ^[٥{[\ J)BEԏ</dVtc.^hK"Y򮿺AXz9+/dijΛ?/mb 67e? W?|/{no~ fW]]F7&:y,oĊV +DY;e.mH{\GaieL0K<pՙ2Nl2dm~DJ`RtUeoe_AmU4k;Sl}kYL N)M#=dMD[UPqgP6~ɱ@Cf\rcUjMLh:)8j:W\`ݸSp5'õ5Maf 0eyvٺH|4&HBIxZ#Y@P `Z {R2:ǘ'F:L̜2L0R `"ژp:EEwJ2T؜6B0 *Z͙L2>}R}sty]YÝ_(`0<藓?9c;j`3Չg+]HL3c%!Ra1$dEig݀ΞpgH\DF#f&,MiM~4K.f[X1x,kY[tYkp./ǚ6m!xL/ sץF08w(I\npՁMh*M!q.4ftnDnrtʛ :b1)PKIzV&sjZ'ar&8Ts*6Zg|x 9SmјO u<rΔ?1Ύ[kem;Y=BxףA|YdVB*pe)Y~D[!>eV3k֧]hex:C`8ց585[ߦpvkԵj#Buw”pS}ݓأwrac^Zi]@Xr%q}o랸6 eΛ-:"AKLlU' nc'-ZZ񵵬iYSma,ljyx Zo8A)BB*AnR+Vx? Iz ̳g`<ް}h<}'' "J`l9 Iأ(G8m's &Iegݘ.@7N̾.N~î4_m^aMޗo b^R=Mr`}V-=u䮮[A~K~;C-Tq9mM6*Zr* &9&Î$oܜGnA#?'](̞(qvfׇF/4lmInƒӜ Wo2{hwHr +OUVc>6b֐Z"+GߐFjlr!ڢSNb+") %Uc&URt){"S֪R$:&#PFwJw3FVgLJ;m#9t]U.i4Gۡ[!c Wr`CRhWYPs.ZJRH Qd myj2F[BBP\ 뚣ʽt3dxFo$VEh:Ee8UTI O)_~BMu XME䤱R:6&R2'NJB /*dqfk[0ד4. ? jT5}XW\Hxm F-&.21% g7؈|l^R++{߾s^ЍMhIM= V&X7 9Tcs\vzh~L ӞҀai04In5" hӂh8y Ke_(MA綝K &I{ڸQFäkaUk _4%&HZB{v(Mé-4ڬ M&`S:0$J qKTl)RH/RUj0g(nB抽;tu)51NgQ$EVH;k1XFjM"U,]*1"Uzr]8˔cx9U\4Rd\K*2 b#֬#Vz"^RZQ |Xn9;'G-nZ'-[+t$AwIL}%漸ubnՅf\0)3<'P.`W/m_>*uQVKQ>οhA_G>de;-/E9X++#k2d[r3\/_}*OTZ"/E!ZD9N^|t!&_ߖ?ȱo&uV~zݿ,Շ:Rq>:Rj_OWgٟJZbNfULν=o4k}{=Գl5ۇm,BCt"3a&)T:dMޅX`QC\&D5'!a$mNȭ71wm t澢"jQŢ4jg_c^a9TCc!7EmEw(!rsAL.8/.YA9y ˒4}eg(ճ/'Ra K;V_k2&;[l'b'b(ΐ8Q֪c{i> s59ç|?u@]Pk-Z]\Rq`8T7p՚yB]u}oXopK%YH x r˜tE|V m ]Tllꃵw6Co^AeVNy󧖀0v|s|,nunCg `].!)ԙoWセRk88!?sΫ<+1 h \s0*vۑ6jj42؎ʐKAژBFgd2ɯX@ӎLdvߛzGۗd g rjCwRgu*<,t%YQҜC| 5G DWasZ9c.[зl+EʢZ>FT#,!Esd![Im8O{8n>fdɇ9Uͼbb{bh- n?._?.MNjYjY0fP9mVD!wD!GfRQh3"g'&*lMlVq3/5YXĬ8-.kHBs Ycc!SAFmp{9v޺o~8tT z;&z&a]sLrLxuOx5F;^s'!zyoP8xRi_zc"#$DA$ygFR%fþ*K0ؔw'6"DbL-+I=U< t =kKS\:08v'Nivӈ1m]6rXer6v`O? iy9Unh&?GHHw*Ŝ1$Y`V^%jV,%\h;Y<ϖ~PO?=frx,BA T2h- Ħ/NU!e+8qj.ʈӸVL..Gd^`@8oorPƒ|Lv1Y+CLN-6lHtNP+O?; SbPM1*1K#*߇ʕ-|vz(SەUp!p$=;/LrƔ6L2:P>>p9;wRTWl ^+_%BKԭV|ы*|_|e;ZSXC(2\,N̺-*!TU,US.MMa˽d*6LdY<,8oMFJmQN. -Gޮq.;O/O :8o|ߟ7`-7x=5;:uȜS&P[VnJJ9_@{qu5ڄۧ-)>85퀮A>T]F3:Q n6+1tS[L>dbB$V1D嬳:3!zG)S8 TLiҡ܅rR*G dzyWJ..Wm_Ne9 _B $ǐ8"TT6MVCVF5EDfcYzKf ?0dSIbI[b-o<]*lk]T@L+DZD g"@޳!'GB|:xΣO64%iCxcV # ^'5fJMmA)h-* >K,\^Ee <} Ogc"{)t< S:nv^.p|sx^_Mqy6mkї{#/Nfnfd;GTE)o&})~#J,^sfMnJ>h3tTk. Q1@~9~PzմX0mBD @n^בnH_pifS$̒Dz-R[9^(-'YJ|S(rpgꖛ57>WWz(Ay:'9g^?.jG(x*hhG c yQVDk8Dܨw^٬_ioS _ߖ߭EFA4Xvو4q&$ F=zyp* :xI$*RvVzAM\y(?߲l&O`T/t]51R1xK͋,*Nw>U=+*C CM D`CLW|΋|P Cχ8Xo}P!1u72$#`0$lw"1 1q۹NscMg  lSv.c֫IUNi-c)yŕ2̥BYay PRzjG ZDWkۢ+Y[ Jt(;>U8` gE\?b-㆟Y1J8,k?2e""/y~7DڎP#h;G֣r-1˷[YnՃ!Wz6BAIty~ d*aoȯT.y@ AJFBd{SggS % 2OF!ΓU/MPXk(-l2rH͋gO׺SA946ʹheya]OE X]8,<>`I#țeZ%D e" Ldu0TSCI +BNWRu=+!mmuhuB`]#]q%m`kB5UT:]!Jtt% MRm+@ŷ?mM8W+i8Etn ]\l[ t(a0tl>])㭡++[3D:]JJ聮4, ` )..p`q>MI/.9 уޝ|ۉ}pET< <"[%F-ܻnƄDҽwAys7$5o;dK@C\58Xgd6JoљϢ6#Lun}S+C$R$'8 (o Eiْ`y5[u!K>Z^OF(H%YXP6=D$Z+mmZ8+E+h(ZOߌ&6]ӵmZiw*\R3Hku36E`A3pihτ!JypG[X"x&n[5+Dkw~h (ahte,٥KMDy>=*xeY˙mU #]iFvAes?ˁNzͦ=BXH!3eIgZ3ƴ&-+ll ]!\BWV]+D)ttř+kd[3I뺗 Q]!] tF-+lc]!ǾLo#]+DÍ$'ɺBbm+D4>ҕ"6zQq.x?ayh&%%< ;2=9z&Gx4.lO~H&v STj\0]들ku³TkB8TD7҃;*W&<0 k* ?}\b\Xzg@x%(xcQ)qh"6Y&hn5x3xrNQ胪áyD>̓E1^TB;:硏 z5"y=Bm?Om۷'9!a/! 놂&+Id 0.3ix*eJN 3bjR{F'GW{k} ~=_z2g-*#.NRVext|/rHٍ>g@kG,닠6٨n'W`OBB ^Z$Lݢt$6b+ıJm*J,)Hc2Kb8q1e_3w>K@gJM"wߋR`I'*N4LjM1rMD31ҫ3i5e*NYjI*iP`Tb5d*.^?ʊK^ O\~\^9mˀ履O*pXS.0؀tۯdtKIE)q7N*Qխ&>ܐ],WRt3tI։ޥsVD2{ԥ eN41 #2OҠs1c1Tp/gtbȼ!MF3 Z/ޭ.+&sWi´LkɳDh4X+6,3V &hw^qb> 0PCzEY s`X8Ri&2Ҁ![;me3AE} _kZGh­IG5wZM$g^ 8N#qR5B؊gAr`Rj5}3a:\=a~hφd`FŐ´V.hKa$Tr C8PFoɆ0Ox/n^DFsq1ڈ gaE/aPzc;ӊsʓ+K.{hG[?5x'@%ybU,uB˜*d ~#WW>SuK}ZoA9z/ TR:Eb| Z9%%D V0. fWCR\>Uѣn_4ՓÃg:i600Fóf]lc-eD|Q#?ʩ:7֜=Uٰ4z6MQ95L}^ZĞG.y/ӻ W/5WFu^U,v)IOf?XXLNY Eh5QMыDŁ߄d9?~}>O ˤb<>:weT-dyt4JDPGb17[Uu튢*7Uy8JsPo^<;u܄#ܦ` i:kCRtJǐ(e-2:U:󴸹~m#v/5Tl HK`^C%j/>Yr5!-Yޝ"]Ah];8h5vªG?NhM2^m4 $I9d1jwϸL볬İOn?#Bw$&R!ؾ`a"WsZ 5 J b0hiz Y#g/4`RClhzBq*:'tנU2mo3Zݳ_xD#َ 0\yl ۊP<)J/zTᣟl׊?.\gåx d 2ș #QV3cUUID2wl _6BXnsUCnfU bvnm<e9GGq2?k_H1_-Ԁ@.G@a)6"kK(H0y@O_Ϲs~ _{_k<6¨PX6&2 Ҙ("`-l@0!S BQJR Ke|9l3d `{n1dὡ%YTF9IB7$t! yS  ,^V^O`#;'m%|  !U?wwf{//vo0|iLڃ} 磷[+uד}]+>Ͼ[*6)z6K"yo9@"ɜ$(w;+U^/te]#;htyj[^ ?:MgK9ca2I^߾㦅̙$aNf>;/NEgr6͢u07g.&c`Rj`j0и9r{ 6bst&%nZt2b?Xdr,$@m1{߶8Ϗp pO7숬VM~ 'qbhhrlǜ5ӏfKI+7 _]8n[-]>YE׌~*;Yn(ɸ]fMَvr%NhkWTSY? ɘ\nnhtetzZ1 Ol' lVi-g3 ><*>鈙Ft| =O;)聉F yb郧ޓJٚ_=6eckݘ^k`|J:තh5r+ቭ..le.l VHZ$p d!49͐0FoHhf$?>hoBe^r' /4:%TmL9\uc__v|L =w*`DNk韥[d^IaS9u5O:| 3V8d撋Qbr"!D=j P;95KݩYIY3_> ;=ʇEE#b*qǢ• nGҳHp)5G"&ǘgFYrA' LDL,ubsgR\KǹIHՖ9,aj8UʲP Ie=o $-oi*e/4rr<_|Y 3 nV"A3T 54$qBgh`{!ER@bAC4FIch[rr p |EabEjWJm,aksRgGk 6C9 ^ine~LsOJKUaVmTPҢLdML,hM)JƬ" ě,sXJ}9Tc_$bqDT%$ o a4 fM^@ԴmyGJ͌f361AҠndE^ I6ke4 r38!ȍLitC>ԮWEWI.בK}uV%ESY.A.rqߎW!t@lεIVDJ#D($.@A.>\<;NXYi DXW я/hVգ'?[fY5W&1` 4^]\*BDLF6Ajh@nȹA"&%dn|@HwУ@HG1H$Y-iP876*3ILEHY(e,bʂYT%Gi3S-P&йled]#!gޅz6GQB^ø];[<쮊XίϼW]=zLhgA:$p-Ia<:aL>{iH~XK :B(N̜@!F0% G0VZ9A<\"Ev]]Q9xVkV+`[2C:7ؑ7جQWͺt^ަWz_giAw}!Uz O_H|(L4o% R j4A| n`m1sUϷ]=>[|U2!7j'Cq)vjLT%`L,(CPJ}ԇROOtt>o3zÜթq{/T0!8@>Qcu v/4ƳͲ/ca@x p6ZDaLڛ=5ioȤTL'ҍX_u ؐg9P ;c& 7Fe!x&pL'֫V8 HSvJcI!rcJL@V""%xeHW#rV8n >nx P (#;i33`B8Í uݞdp RZ.R]wg;%`Qz͜Q2j&2FgR[;.l|]uìw,ˆĜ2&yQAb<(d)"1 ʓvc籘B # &τMH$AE̊ YjdxbGxG<6Kn^}%؅%p1h098=ɴcj:} s4xR+fE|r(jzbj hE΄-eI|ȧ@"@GֲΗ2х](_E\+.] oL\Ͳinm܈V6[Ol|RTr՝t_hMAY&u{t~-ßn-h ѻoF2֤^ϦI.M<~_"*{1_Igg]7*y-d!ch oJ#SN o/0=Zy-$Z dIAs0eSP;:)Xcw>S̏8XbYb7tPX{d؜zw_I(v;| ުzl}h}rͷ -ض|o2kuz!eOa˴FY)Ҫ^6R9Ao$l㕷M, A]oWljz8a ݛ*,.c*Y8QZY#ڶJ9}49JsQg#YMNi9%UtĹۂoY -L΅lQ֎-`Ђ+uWX_ОDGr"N=hJ9`P c"@n.Q>r<Q:чN:eCJ䦫6%tmqs~2Yõ\~v1çXm*J43Fh^&0eu՗BJH]^U˦Fh;\ո CA,Al(9^( \%ɘ+^+@x*mONf[WRŴ钓mHN2A8H6r2hrSF1Fֆhfk:(,*W& O[˹IZƴGSZ0Iߡ騌ĹW mSH_$%%{"OLR *7|.RKY*iQGic4:M$+ ^d/'*DYns[]u_wvU&ictcVHdr#Lib`\LFAvBs%3RČԞ+ MQp[rf@< Q1mnK0vQ1d!XNF# lLI8k#lKNt"n@T44:Xz18E: R`SÑCoj:?iGr2m㥭τekE>+*zWZD~!pƁ p44?%gc:Ѷt[P}iřr;:c-'+<ӌXXW,Y[s3"$4YEzkLEk|a[ey`1z '"Qf|H MDDh=AI0l;A^ѳ.h;l{PQT.UJ/3ҋVJtVn{"%]7X) ?xi ٿl?/e(fe@J&'*/];oomJQ's6$lݳ'HJ:s\։|"id1zH) q hC>2 b:e,eҔ⨵% q0PNRxss>aԁmQrxlݔ%)A=s*FOt Oz<㟾eFe&G,8!eJ4DYJe_/&PKTbBo#ƻt6_[^NJժ sZ41sc^h:4ZsV_?~Ԁ?H!VJKdd22 ʖ)tU;{0xҘ_+) >U]ͻ&(WA[!e1h RHbbY%].Z6bu?G=()+U^pog3|ɧA]Sd= Ƴ/J¬Yk|r&g9xdPFo fa5A#2jiyu\irC'Amt g΃ZDRl+S`@QfbDL@R}NdHkd"mf0GGKٌ|8FQ9݆2;YmV <\MzRq9xtu*Eʧ  :ixqg#tڐ'&Fx£4?ל'K!)S0ɪ)|h?W,+:iy ]waXJu71r 8_9vAe ^FUvnN);rx&A *Yzg2`MBAin!KMx1e~J5+R!:E&+,JQ%#"WaIAlgɋsCyX_+%M&A戮~D|u/{@IDP~+1tGq0Be5 _¨޿oim^t! U dMbSv2bȮZ.2QI7չ<)rqDqu>Lʵzns1M_."/wtQwqQpxeEXҢ`vyQy_nq 2vCcnH?O"pg?6ԟ2(]ro3]72 [꜇?~[r:+5w`]ZQ{n0t8#8LL=\{Q~{48:)l2E֨4;y=f$R'&|r_Ӽ{?NkJ?OJ<ن?pǿدoуouf$.R$X%ſ绤.c=qf37=;PGRiSO%9ndžo6[][f-f_ q=JVAu(@{ Onvx<+IϜ[4yB^rOT\͓|ݐtѦ: ܪ83#"x",upDKgugfzRtpi=$] ȏnYdp+%u1ǘ2;'![锵r>dR6[d3Wy UnJGx@]`Uōt+< Oxi2ɴU'>wLlu2]ĨD`s5Zi1)"(4SAd4A b"'DX"dD6KJJ ca !@74&jmإ[;P?_YXoy6ܶ>pj}Ʀ!s%L<ϫo- ѫ@DZ+

Gaۊ6D^J͌%f(Jl!gd0939t^k\ H@ZTNG( P2EJ(yuQ9m䘒1,`B aJט8wSej2—6E&uݢ Qdl|: iCJUN8|2o8}9t~$3 I4)C [N4_3iIЁg2!fڅfA5s)1! Rgh@%,&V3RШ-B Jcܳru YJ{,Y!,LA)"q 1L^( *.JLd.AM uì4֐Ɇ?{WƑ qv_fdo!1n+dHJ_kɓؖiTU?U]9)-8@fEBɘP|]є}S1fTL9H?$.iB&lBUky['d3Y2dyb)\{Q޶M/Ahg mLɠ!B@mI.7yAׁJi@ &M#$F }ti0"xҀi@۷4I>Io" (hB4N;"m V5ChuoߗVЛNߒL]ĝe=;v&64lC'u2l(G6NrT7xj+.rxʍ7AdEJ܀꺩Xpչ׀ QKQꈠ1U"&d0MuYs3^CYqZuM{0Ll;+M%2k]ItY7Wm.{ ?@z|j% Y+%[ѵo:3/' /raPm|lԪ tI r2䋖=*C~Sd#!^If > ]>%oW!dV MբAADctZJYr0M&75Kdd+}TUZ/fsz{}0_7Bܼfګ۞Е}d=f3}mZ> h)J+<ق6Gwނh3%5rR-ۊ@!I.y!].' @F#Ĺ߂oYo*wUO^{ o@hfGLwCd2eΰ+&w{<^wdY#kQ2@naR). gB)T=!eBHգR"dN{)~ O!JxH0YZh@")G&( @};& !ב\|!5@i 6r-XB#seZM/ rAՇϕrg4 ﰛkؐcXy7 `$ f/W; ]l 8.â=RA(K9 <љ8'2Eހ8^[r<͓BL :FZ&pq4XB )Wq )%sRĝADV LE\v)YgLH8֮>3tCQp_(RsЌp sZ;U8'L,/E` Z1d)ybF!ş̂jy_gQ?Ot"Pjme t!f| %L&鴒6TrۆZq5!Nz%TxY2S05 ! 16G;`&Xټ/toπ#LFp#qr )J[C[0R!Xq̄Wdm^Fd;[$rЋ^SV6m#.8Wcr9uh.Ĥd٨5 Ax%3$ +kl/QMCjuo1 z,kAءdr"ҴC͐K[e[0PB.[)UZ'BaGy "FIZ5%=Tѿ ȵ.)gc*^/3d>{ @}x)䕑mwLz!HeݾRnis[RJn3Q dp&,}J=J䘞T =wAQ9D͝ҙM2O#L)9^ Μ`,Yld*KM!H{AkVi) \Q nATNUdE#4=.yΧ-ȸpΗ 7sn)N?G5]F J1iɉl+PK(+m,J$׶JV1 z{ i CIΠR8tos^tnC|BWpE\%_K"}o8[|h8 6Nachrnvng;WSX>~1R*(q* /)t1t;V;aјgm \&}}kz3q.LK!:zK>YR卂C.Pds>:R8\"}LyI8_m]5?=Evi:ϾoZ21@@$oq6t 80&hi0eAhPtowHw,D J$\Bd¸BУs :\$L1IBϦ 3=ޤٔO0.tFwxGdOEѿʧ5|? .:dےGX҇O?w/ gѦrnl+綊l#릚l6mJ!i<\ȳI闋k-ߏ~\ŕQq]tb?. CZ_7;Uo(yQpr'8oz벤uR2:N}yĥ|7K?v /0i$ 5x濤%Gc\߳ü[+j9[x)v>%ov&|E[fS5˨89'=eѭټ}:ޏ::^ђx~ȍ'49wzC=I-鶦))d<ۓK71_V8+V~$'HGy&cA9{'}DJ_ޗѣ l~OE~=cLx>$-/˗QlO1::;F$a6 xɠh$b"~Ɠ_~gJ=Yta=5i{:bkGiP%ſy //}nz:lZ@[]6'"LJs\ _t;C;fgl;>M,`Jmk귳w7n$o:ڣ.8 j˓f_Hܣ ERY*.A]ӜM{om`O,&8EwH VwFv6]sΫcw[/.FDghr(tV[uz_s23ܰ\[(>qm`{GNDN*Q#1v+qpsGpX`j :9oX{Hrd tʃr4UR˞WSTiaBq7x[B:5@̒[lhÜɔb.sH"21UqzEc>nz__M7DL$ 8ߓ8!*2_뤂([P?xzk뵔k̴4G(bRMi %&Ю|{i,EU?Io]_! g1g `Hs]moH+vhKuW`/\ &3X`_m]l#ٙx߯dٱlYm1f$H駊Uod.$ְ>HeH!w)q!XpPR8TI$EZ|Ds,B5,ڂֿnSA6|eo^-W3*ӢsJVƉH]Qc0880JjR4iy!|I)sbXKk|[b@,!F h l,` r4#@*3Amnz@roc6Lu?J&R|Ti:X,!SQ)CBzz!/?(JhiAZ@( 0+jCTLʜ.G"XH2Ov'Fb3u<DҚ9,$ P$SJ= +VD vV 7 ߣ,e%1FIcd(>.iA8&dx,ξA4;k0'pGMJ{ٲKh&stȍM{FUwjb9 w;+U^ey6C{H[4F'G @@"xBΠHL2>4]2Co`00(O/>>,2<>ҢS.}>M'a@jtl{eq*Ys|z~z8q1p J*@*ch$3sJ:&k'Wp>ϱ73~6eD2#{F5ޔq :/r&+)!oDm 4CgA%!mHnF.#ȅg<@F2Ȓ! |h;ySkpf?u'^#(ٔe^=/xۈWH! C y\c%2i:ДH¡BCtJ-|ؔM|a#PXݝx M9G?>Sfg|u΢8u F TNZPVAj!W^\IJ(\8;R - Ht#2[' 8Fbd 0ZRfElw' &c,d!eɍeVЙT "9#7sٖukp xE= yOR$hoOpu~]]nNWcYut1MD9 Y'kI 3̳g/EpCyKv9C% G#,'ZZ9Ass\";.58wOUYo`2ٱ)pU|Փأ7bcs55}gVWmp]nmg]QqL:}P󠑰ŀEdxK,{I=v_vS^O޹LM1NƓ=:c~*0?aۇ]ʘ(hߏjCa(RtFF$d.- <+p&$nf?W6}P4»D蓫ّ4Cg^W85hYB hAa!f\G_Asّj'4>Dd4go8hPzV2%FcfRs&fwEMKע8v&8*fC%JY皚eDgrFd|֜Jc(;*+3WْHܨK4Mt+MWhUbiԼoEIFחv7fI :un‚tPqFGie!8t#0]Bj x*W*, PMNN7t:>Ng(^UFy@0~҇x?e\.{+։ӋQ]n04<=f-otz7˕~WY$Qʢ*аs*YN=;N`r.zYec8 WR#<&2>T/`sJ{E|Fx_gF)Iłj]8?t=7 ]`b[tU ] J4=]@FiUg+tEhmR^ ]IT;CW`]Vn*(WW/CtU+v,ho;](mOW/ E1 'zy]Fm;]Vt骔=]vծb3ꪠ[OW=]@BMꎺ*pt@RK+cCtE ׶] J^ ]ZnNpתV߼|1Ǜ P4}?}vB+~J=h4qs5{TtC/ъ󐇑KZ &ڹzBӶ)o\yo޼?!] M{A=jDo{ edyﻟpv3vOܫ]kVGھ7o'Od;oKeGy"213p@5xgͧw3S>1~ :C]9һH`tֵ\?߾Mƺ[2z(Лvj]H39].$Nx)?yۿC2x aWhonOnZ}6_.k=\NRQwW#гn*dTqv,(Fp}}49e֡1{.ԩTrs5Uu|btlm VOw煆ɕZٷ*BE«ޏh2dRWY-Z3VWַ͉ZCB:V1Jk`DܴA3j]S:"hٸI@14%$T,nmِ*jm-e()oNBcVU˝ֽ<RIIbUTw$ף(]_^ܷu|Xkq"&q -ڒBGZG7H }m f>% H/*M(- V\jH) AFUb}ɈX8pDڌ^QUɓY{JtAQ]|=u50sS"0f;DrNU.Y{V+ RAvT4JkCvYwiJ2U(_ "M ]d0`g`jb |SZ\SQVaD9qnB!Vˮė BLtp EWXLĭd B2fCL:vM\s \Z(˸$q{MƬC670*T)JvdZ[E(L)S` `; n+)jWH JjV\M `6+(\dd Źtҝ2_%Ce:*UHPcفd2 VS 4+dW$1w\֕!ՠPwVrE 2nP(SP|k(p\`N!a@YP6{i]m1y}gu9ݴl\ׂ{UՂY- mCka&a-A7 /Ml#TYǪd 9{H\!QUF2bx蘊'8$;Zg:#.c) >@E&rZed^S0%JC6fd8 =h^= KH`Ge YNDx$ۑ ox1JrA#P`ZDbYx¶0VA;> Ddm~su~hT>COeLl2fXbM= %DA>h۠OQ %:]%@_>1g궣D(ʠv%P2vڱ", "+P(vm!Y5MH#{ˎ:V٢`8JA"cQ:fWc4XdqYIb&֪P] C VA;55\F? ʰ*)᠛FeV]!ۖ%Vb|؀?X=VHgYtg&)hP*ݢrܵުjQEx`!-##Z6Hª(j` F{ŷ,81dkv6i@={y__t=[—|=M{{vuFe,aUA[.n[CiA'EEuh0N ೓*"bjЭq5M/ yHY=yh4vM ƴyBoGFзUfĞL:5n(!/[tlM1WdsC<܈ܢ4D+YLw*TP=`(uFA@6%fdi 3 C)wkiyf-v+P!>cw XW \N1֮ur3pMr,RDyŰ*aSg(eQA87F1 w01t\:'X:W!CڨlJecT-PFI6mLܬxZ35֬Y 6jP%h9ub1KCBfj)BvrzgiyנB݄hƒ?j'l5^5pgB8ڢ* g`*ga2SEiaV3 _B\8/O34%0d*AQEIJrlCɵ+F @:h ` Yc6Ք rri"&b9\Y.8IP.X4wR@Ւ0[ Y\ yFB#S`j?orٯ>WWVްjKkR]lS1ZzQÖw(N/Ei̩ֈºƕuJm//z3#VIU)w7 M%?]N_}x'38 䭳J@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $Nun_i[ɺV3(ߕˀ՟wݖH޵JSKVPnus .=zJT (b|/hPK2@ 1@_ h9x[! %(1@J Pb%(1@J Pb%(1@J Pb%(1@J Pb%(1@J Pb%(1@J Pchk=an?ޛV[MG?~.ݾU| $"Ju&RZbKq9xP ƥߍq9tt@OþM5_뫑.xxwYOQűJ uq_}Pwxܮku]^\8Cq CgoޕqlٿBAZ}!Adbx *"9\؁ҒHK6;-է=uV]9: * O*n1Rvt7br%2u&7~r KsFkKiVMdȘ8ɣ"Ʌt\n۝t玮ٴ)uDLlƌLb b cMeN9;~UFVȼPl2 h+zq1tJJh;]JdKW'HW,DW樫6F]%tP C/i]\xc*%)trTwJ(%n#q*1tjH5NFLnR{+DS*ոt(n)ҕLƴ99o$UWHW Kd*V͡W7f JNW %V-] ]ildcCW9ƨҺUB7'IWR3"kIu(]ϘVjMlv3%/>+(B_7Mǣ~|}~|.YݛmۭYEH(Y oΜRӉo2&5^xa?,⣬LqXt߈CW޹z=mi<yf1crVoF %wر_nhővR,֗ Y+ztP +,k ]\TS ;]J[:E"D#DW nUBq*$+JɫTJW)#RѦU՝eY`tu:t(5 b1tƨVUB)EKW'HWj,4_cwUt~AtPV]"] +!1tphdrVv@)Z:I,eh]%J4.mΛ֞JNOS4i˫ *gXDW -=]%AKWCW)& C4G]%J6R2$tutS;hIź19`nIŷl7OQ'h9,.Qݻ6÷B@PniJFz&:مj4&uc(20ûv+oVOY w:6 F#GOPP)}j*<)? _edp`8{>??enflKJ"92Sw7yvN]77|_&ۤܺY Dg K], E/EμUMK6^DֻІ~ťk C]&oxג<\^y7}OƿIцi5~sH(\5B0YiE%+ e(Բpl=?H#RoDEnŒUD7.$F++40@nm/ LG"K|*+Y^\ǩ˯םY? V؇e';_ xgz e4dLɿtg]LW2( };5hƟޭ:krJ0^VW*7z;{ot@0w.pPj ^rnLHg؉ Ls3$ѸuJpwt{M70r_&Ÿ0}(`](`2QoGZEKۖ>~VbvաB@rǗ7<#Dwy󫦘kaa~IgLN<&d7/nwB+ȑ %|7!}w+khS|E @ THfL"1!H"bV !1Bδ @kz-.ОXNb:1*Ċ N̰$(1Jy-yE-2aD+"iJJpp^t_BJAY&$-k2#$ /gep^&ӂbƗ^~Lh> @j.}M~㆛,.r4%P}rՋ|\_Q]̫z~քNsB 3N F:E,dz[F;l~>fͱy@6. u{lssv(l) h؃b󐴨h"}L> پ+ˡ)UHɀ1T)b W68R$#𓳠eRz\KnD&i@`*epJ;~p9I#5ܨs)K#=п$:t~rs#/`x?ؠ/k#;[fW {S^ԘpaQrulp> n'C'B=nт ̨A)lBU_'2I.V71vɶ bK*ɮ6Ȥ5hJ[IuIRz)k?ƟV Bgg `hf)1H)h[)u*1ύ=eB4ب8U:^%et`{!E#A(HϹvJ˰d =!"}:⥤D@XYh4TXVˀC ڼ#}Gm=R@ߏ&m2"cIv?-|]T@7yU ng\`g(c! 891a,a^.mE$Mr|y+q޹z|wg7&ےAEӡ9T샙}o4@Ѹ X{Vz˲0WMAiLH21 1CcrbDؕyW^uW&F,(/'0~5W ڨgPV9!Im, QyWR#F>-0"`y56V* REk#^̯AMv t8.=)1靃9_kk/E_Q"{)כ/GqYcy)pPQ'CI:<cf8L -^!O3^*q-֙r]QYEZ#oZwSwuS@ؿt<rK*RwU#N:c>$n<4o6ޚ"'ݿ{ߨY27)wpӼ"祟 FyŗN"Ep;x$I(e :צ7No\F],@\g. BpY+)F ôԍ K.c/_kw,xBev/\Ι@$̊/"gQ7?^qq긼lX`eJ:1DX)2>^a6΍goS<ن?p?\nϧ'U$D)/J1$WWxYVQg,yfm =<2 p|,֟<ב0eɆ1vtwJR]?GI>%MmB1 jl&?_cK ҁq8Mx&7o ge٠hmRy0ׯ a>9-+g'Ü|XZ0ѸwW'3yWC:fq̆4/y/ݼP946L 3Q4Oh1ٟ0bg۩v!@TѮ컶SXS9~Ymi_=S$4H΃)cMTGԚhT˄fvIbSTы Nx}!Ty#87Ed2 y)b!« !-UBz]ju teש@S7ɺB]X{QrV?(Mk۽K  Q^GRZUWD~fJn1\fCYʲB(cZWiLI}$2S& +`Zu#DF FZı$p3*AqZm (W$6!`BLƒcǬQFYJy>% nKȳ Ζ[lZ GdH´{C<Q8guScEU'lZA>\Di~lϴ^> &wkg΋$9P2BI!y;{j-uy>G{SV{5so"{`*9, @7@$7_9І.MSCHk|hm}T?щoUbxfkuCspB{6+CgqeVykTd(:u%N_9>{8r+GUbCu:.8i^]6,[Y{Q=9v>oB\޴7mѾ}3Iߎ#//4 U&Q29Ύ[7Jk2oIwz',$}݋o_~W^o{^*z^^ 4Dd M_E ࣡ڊ5m!z1b﯏+`xZR* ~ j;I=hhR_)@,jmrAaaҗ xbO5hIS1+߷з#aXBR\(@]s167&P҂w $UEu٪@x3<ݝOK;7]ӄXR-Cwgd"kF.2} ڟϟu9^~Ttv~?/=ltfn3*ͫoEa+ߏJf墥 ͔QF!G֪rq[ƺ<;„}m|>Pܗ`ȾՄ|& ]\z5grMrZ}|*#V7:~ew7?Xysta<_TvtҞ&N_$nzzf2no8:gklIj6~kR؅c>[7454!od2jvm\vm)-e((5B20t$̡(f`Q ۻo?19{M!}&->ȜGyfl"p/u8?;4S gG|zո k!M"7.8$dx^vK^މbL^(TRE1:- HCƜ}EIs uHч$@Eޙ`1(;0!rJb[*M,18$? Sb@$!˄!uouuC3JnL!ādƓ8=97z,p~d̋NEvl>۟b@a>r8PiV|KƊ@ dޘ҆ޅ/W7? G?2 y~7ZLKx:]XX|`.USc&cze~l!;J]_q¡?rϻ ӯqt]eBBKƸ%?QBZQWȆQqN{1;;2a؊2==2僚h&NNgmݮX7U~6B{ߕR]yݶ6,bbnyK¿rJ7߽FZҺF1Wfs|Tm}5HoD͏o˷]7!ʔwtz~Mkן\yyVȷK PW7ܸ[.U}:헞zM’5/m::xQ=8w޶G-Cu]!xK2L (c-s]d0Z\W]A T|H2Ӱa ЩZzF,R iW^E3Th"60aCYPYɰ4Ϯ!#s0>K6meZys-$B"ʕ"mtAI!*@ޘdyit5SY’R AP*Z*V=9R$01`-'GG_\c{X1uqǜ3c*0Ra1 b)k␲aIiƞOJVUmjՂ%,l rEo%2HCEftܚECq]6 awvG ƀRhkIHEW\ 1eb0V:ZXyP}XLd)4U3IYDh:Z%UXLe33٬[o]Xh N#4xS2!A㩘 2˻QF+Њ(E bm0lsKGO-](uBPHh3Ȗ%Dе#gF0}6:㒻EX/ڝ^uo"0J@lEl b.-fgEfŁE"+ӋEwՇ4>cû?m*q].T$~߷^4W'Ap LJ'!]*eM@(MP4dsl!&RvB*uEnAHG(2)|:ڌIU1kօJf"+l:FhA!hI!e_: ֧VyPٓΒQE`2r|lNfη(Q黗9i\{WW|*1T?ŷ,lA$)PFT9l ƺ&ð̒A1%')'}5^A`B*CWBD]0 TsKWŇ\hCy_~?jI억`j얝D+(@z"*D"9ZMh̸Y'*F">O} jhe{ta18t= h#O94 ClHXL5Nsi/{nK{p\lKNFOV仕p JX{/Sa g|@4E@etd܀Yj3=)@q>YNcH$EI9 ##Og֨D!CȹE1w2=K v"/xTdT(2lHQr (80,$bTEȣG>[^Av="N%iT]!EHM2: F',TanlnxBmcE6f R}P}b!")Ox_X3h\'qd pK'$1h06aZk[R53+t2,rH-Yu_q[-k_^]\'p3h|"wo OŘ~; ;#WއZ'r?*paS2+4SVGYr.WYz~=;<ϾZ85Eefou#IB^u歾K&y9-_QύxnuC+c\ݡTQwř7mD?@.sz6~W.?Qv1^g17v=zROf_X/#_?=^J=|&:1ЮzH(b dSt6Ve|] ] ܶ] gN|Zd,D+C\-j0LHEGPMKNu?;'vMw9`UDWr$jHZuD!mk`=j!I'I%o* T}Y,DAG/m&! T,xJwH_> f7 $"|ں%G3?Ӷ$kdږgؒEdb>C6賩E8;A:OqNh6ɬq f u@úlR4`Y ,'8˜w1&K]I3MEF8fUQ@VJ;( ttn0Cɑ[LE/UX!ii/GC>cޕj䠼+JyW ОiHdNTž;RC*~ 74Ҵᛓ^YBRUu17嬭1j=oӊ/?=] Uf>~8z]#/]nKv;DV҈{$Y<0Xf㎬ 4zQ1e ؈=eYoU6qp7'eF8> AluO< [bT~! z7MjA%ۣ=? G[G}ka'zs5lK҅ŝ@}`0m k ͞='km_l;LHˢ]hNF@Ys<^chAU@]~&q f6Y`I)uA]w֋\uKd%5)H ѡV0E bIz!&9s,>qqZe֪/fmϭ}MzjM>>~Ȩ-b͍ݶX;9-H\xG@@4M"g5 N֥_݃ Y亗s!6ml3'a?aS("]Mg泾pZZ=h[ZZg8}9kJDt.l 8mfŐX7"X^ +;+kwI3bparr\k!0(hRjVg}x<Mp;rl)f$.β|&˒`7ZUL?_&~LjRZY41ʊI.\ʦ΄|$9Z<:8Zqp0{'z'2D, 9('#)d6Rf'l81K5/$Fg}}()d~XK\/ r son1‡4ۀl# \~JﺃU` co6:3w[aF/(^05-Ÿ+t7^&d耦xzj<'{ޓq {|s*nd7. x_d<_УT7(Î-PXڲ:8yq|Hq$@:'hR2f+H[ҡ3.dP쒚Xco$/|8z|i:؛6}Lg2a%]2-5{1^JE3-)QtLW ,|1ꪈKQWE^Ru+TW9_]^N"—")睺 Օcbҍ'eXVWeۙ>,66}kOJʄb4"D%o0o1ח|K;|+܁:m@w= ?F?͛Lܶkw6t1O)7i6 _fh|fyt/n BoA@ C8(2:Ӆ ޻XP\C .D7t||Qk&W͈o_С J8 ex,˷Ҵ2e5B{U9*BΚ9%Ӟ!^$Q.W}M$ODwj{[ˍ-:ڪ8'!ctF:R%Qyᴀe6!** KX0h#4u9X`.{'G!)ehPq~0̙.ɺג (Gd7Q75LY6l~✖4>]ARhTDo[7ϳ+??x 0SYJ5 OW6.vcjmo psM۪۫Y巫Y*h4:zbYG':Vv׺hZ]Z'%0-fHiu4 i+o fw l~X꛶X)~0?,x"x$LĹaWzGlK?ytkjTE#&˽пe].mU݁g9?2,ʀuijiA&XgLɸ+@0KŽA<nYX>k.5!MAGGΐ "7ςn0hXU5ˮ*Nz4>gg$Ǔ'V خO1`ݛ;%Ǐ<踑;űCGφ/҆sਈ /8wVc?pT8Y!Dq8ey.nA1?OrC@,0p)t*5; 肎.(8ǕC1,HkKHɔ#* 1+]֥2/ f*CV sk Yx edjo̒nnh/0X; ISܧ9%j~֫7Gg?E)pX9g)mýc hG݀d) s/gAK/>:+t2 @/]1cA:d[2c\ʄB!H uݳ:%, 3pRyo9@"Μ w;+E2&N3%[jcTw> $i6!796-dtop7yO~df05 ? wW-4[hy]i)rzt&דqX,Pţ3>O,N-T'8m=ӫ˳mKmRcb d Xln_nfw.s8=$uݽHw6tg3{[{vPa;v}<)hjs1E1b 򊗤 nGHpåC`n"et1{ :ZL̒ R=)`PLĒ5gp:Eŕt#LJf&V Cu E U͙^dzH{:'ד0?g74n0}'/\c;a0d&u٢ )43 OH"cIfBݠb#3P7cEB6ҁԂ:mnjl9MZܯhLcFǡZ*kmv`w:\0xLXsmFι,0xT:4LqOFKUaƖ YeYA&M&$ S*YRٍ# 47YYV~}%SQǢC5"VֈiN#7A?{WGk_%/P;-ݯp埄/ d&Oa4nKBpUIZpHb7PX,^M>V@6@C6 +XehDlzTTn\ J@z@ƪW9geK`G6#ϯUh\\Ssl藸ŧ,2*1N2"Gh͍5zJ%Mk@K\E\K<fC,AX}"qOՏH4qZF}:SXy׳1-5);l&p%KRBH.vv8BHGK0leL"!J@˾b\\f؝V+]3ڀ԰6gXJMM!mth}xmB[j)5ἰ\ROkFUmQ4|ts>ݟ̡^:bPEbrhSj W.Ul-%Z a/+0(p>b!Fd 9)Y1A ٖФ<[հ<ķk- FuwhS}٣wractuqЮݪem 9ˊ,uOܕ\{MDS:OZ%OM@&Ib'hU ƅFH#F|k!fBP.+ϰNE\y$;ң8=d :-G}$Kz§ ̳x{ÅG&{x|\ϞO 4a8 vf7>]*sg=TH"e &߼-$D~w)[MЙo%gOJ{lF|fvbn$ZȽi|J4rDK.0JU`WxThl=C% {b*>BTdr1J@&`d B.kCay^ub(u'Ipld)5 @oYUٴVCM](LHy쮀,2O1ѹh=@]0PM"N<@x ΫSj9>) ?Po8̍Ųl(d-.q/V؆D<[-m-QQgH:мF_ c$҈E@6LJ)6w'Ylά`*T)،|澮8W}ޒЍ)MhgH-} `spE?v}~ ݬ^\Ӥ71h"$/?YWE:'@[=ZZZ87U!V-Q&\ LLa92ZԲ$fRY22t!3Jbs](uY(_CîĶ$ۆ: oHꖈL]5͉)Sp39J\ V͡H]ć :ȒI@"dPZ >Vj1u=d!j7Yh,β gYR:%+a byP%f=s+&RuhA' |ZҞ>Ǿ+ d+)(x' sު'3:~"U(}9[|WV“Zؗz9YsQL90E?TڦV"lzƪU}CYۀO?NV5OwsD9MJ}h\jMdC$} (# &)ːwYb,1+(Ą> F\ l0q1pI}pZp|OUϯd%Lާʟ:#&~ᔿ_?;vrdcLi{ ~zXo-~m7GA=nܭ}݁m홝[w&<.HkOH"y3{qz\oPZ#n71 bHqR/P]hۄ$$qjzcO+ZlA]unX\!`T%Q(xW2&SĂOf#O.FkZ[;Y{mޠZOzeԆa{ޭo]h~z{z o=fw+zzI "z(n?aybG8I+J P,o7 vwuGmuvT׷XwsX)|6$sB[U*ԫѣAK 8'7Vbi;[+}+uΏփW;EX/fx;\1/j9>xB_eX[ȶ:[ lB-T|c1$l4SD:M3W܍1&5RNbjqj2!XKٻN2)6=Qm*G}`C}VqNq-ⲙW,rOVX헣>MNӯ?/FPH"f'Y4vhljǺPQșk7; 9#3`_T! F&TKe\P:{-;0chlb38Js)Y먤(Ra,]~8j8u\ `xz+}iXNzqI(=[XiWaC`{0$wÂWW+!+yۊ5C9|t9W# %I\MSH8de_YO7ܺ+}zm]}_VkJ$[nr૭h]Ri.W(Zwcfe^u[xuOMztw ِJr]`5b@%%uSjo'8B%E8?4fM̭i?n#SNPBDo;RA>:ܻ;f޾Qޑ XaxH/k*@(GHzo'n&@ Ӓzt|qv<#%}d/ j,V|œ.0>Yv٦\b1vdnl !{\|"\RY"@6؊ј _sgp=)47~4~E/+~6}/?Q"+\[pQ/n&Nq>"PkÂw}n)'-IPR{]\ jZ!#EX0X0Y\*J6j0>mhBs]q E7φ ~_Vz)CշKф@q 0b6]P^@M:CsT4Xuq.p5#'x -G ʃ|$71SrNgْPccC|y'> i݌tYw"=dZ2ys+ kn ڸE@ϴɶQވq6(th{)!8KYOɦ l֒m dٰ<,Әx~z:–j蓫_㹆6?:You~Eo'Ga$orrӪfoD]Z|} UEk|2^_*?6!Ϥ`0=s{L|=GΏ7jӝU_=/>㝫DL> LXk&'EbiG+7oxjه@m\ s:SzmlS.D$(TMrh'=N+sΰA8>qOWJ};Z 9=J3v|)٣/&F<,lTz!2zdRi%DDnsYzGd ?0۬͒+jٸ$F}zʡⲉM0gROV%6$v"Ex|Y?ҫ E v~{lءG<}u>nÈi &5x2 /Iާd ][oG+=%{2R/vN@v78ƂP}I.IqS=CRDJ"5*qF3U_Uq cPAHEP;ӌMfS=[UOy >aAȟi"sP(<(⤡LTi~hRw͓:Akc8ybqR)"^(I5s 96 /t 9 Y_@$Dʱ a][d ]DnUOdCx %x~DinA={jV*+xJQhq2Pz;po@4\;^Hʓ:5'85 2sS^F-R[H@p҂q>s5Q9nN4L4=WԔ@̡'\ G19*Q(v7Ew90_~x:6jb(p9CI|tpNe%^&/QqЉqJ*bsS]F\dP@^#J/sOe-7{UDKh|^ ֖\fm[f( N.Oi -W^MVRr1t| {Vc;"xR\ٓGx?޾jsPpgdIiItA{EJC4+?qK+[v0QCiw:SxٴV\zuL6j/gAI~_n_2K^1N 06X!LqNG⣲H'DB<|P; ǛW Rt9W VGyguPy/Üx4Xoo^ݫCEnb*@HE*'tY\tɜhr%fPq7|bاQ@ݨ]PQU]zj U&XQW\E]ejmUE TW"H3`ELcUVʶLKTWJkB7p72 vuTSW/P] k8$c p|W\y0*S[2CW/Q]JH]e9uRv0`*S:ՋTW"G,~0:ΪXb ifKX|;r\d[Hɳ8{usa2XJV7}_G??}zXN}FpT|:zhpW!SsjEIq뗡y4/'ּi]S?po~8z[+eś>Go`gCoՓ18UX~q).-'Wl%e'<{ \u?Zo3mX1nzӋ,)nrR!ȴ/ulPEͽDrӔxˬ [T<,'fnv֔X gix2h/]Ӭf1=}îYD]V)ڹcfQ*u8NZ?/ujuVhП%o[׹νrYj~a ?7[J >8Džʍ#DDL əcSF,7Jp8$ܫ͚ٹ5K;`v&@eޖMv̦1Ӹ s4,6O0^l>rͶ{~'HJ o Yq/A,-(\y2Hn-IA J<<7:a1DRJ`$hÅ_#n5[G Ȯj)+޾LIQE)ˋu;e}c^S3#K $D1Prמ%!Рl05)c$pqwƢ܂*D&Fp&)XDò9oBNR#)rnY*+HBG($zkSA1 *=IA/_P\muQg}EGRI)1g 'hiÌNhIfg.*1|./\BF$KSmT%&^5GmC"B5$@'cO*cs˶"-kkm) me-Π4@MUq܁ӹarBCn@ DUxDbO-w궹9op/Z^A8&&25t!q/_a12J=ʞ9.I u ekna8q:PM:x̐kAD+IZ=>JW h"tJT#<#rɭ5afM{0+oÛ!=igD ]g2rz[|V=v=ԓTQ|1(eB\7\|^tD1OnUh}iKA3g5+?H(ks2G׌=R^o;ӷ2A'+8Q7 &YU ZUAViӄd!Cp>ODv5o_զZ/LC sMCxNIJ:˥A%CКD 9. h9@Re-P\Vz0`%I֌I8P%OMsoY@',m\^;XziKf24o+:,K%э[*Fu{ޟ&_NPOi&$V*\_D%K,+ 2HIؙdBu&h )ZaB#-h?:Q`b>$O^DLjl0bV&`6Pc(@G3!kq @|g'@YH hcܣvo9~sq+Y 0b}1=ݷ.}ma:IVALWj\ U%nnH }cHw<jB=tA(d̥PTT9a%Aq)Fv؈Nrܖxm .}~96]OdtY?]m,*jAYK B)V$SI*w$:j#}uM`Ltն^IUpH9!Z9%Nfg)͖N{eh@*S&rYB,cT;)97G9wCdAF*rOH[j qם`_n~Pwk@ &NkSdRF9PSO;VU\=OVvۜ> $>$$ΑDr%G3>Q]5&0 B5 Q2EF҄%,3QkK!$6J Xް5F=嬒?U::Qr[&' @b4P- n;3G"@aHgz8_ܗx~؀sX6 B)Cʖ [= I-(jdQ8I=U_Wa^2tDM8"_8s鑗A +oi ;hx!Lh2PN}~nUM9;'=XtF['xm o3ҦLJܦ\2H8Zz\O;mExkGݩۭgS>NZJ䆧UtsN=wҹB7ta=_7_ޣ!4p|.$C9抅|R;%^Ew0(7NF}̽i ?*'7V8E!*Jj(' #K f7^cc*hMr^r~ 'ےul*715qZAY"EzfL[̣1 υ,68҈xdz;Y =hjo OASM(DR[!(z\D( S$R:/^3/Q 0K^(aL&L~*_I^?_G*"gPz5-MR\fc`ÅLjG -^P`{89MtP%wSVJ񊳊RyƇLxD3LjŖ]F˒&𸦵@ KG[P"&48t?ۛ4vE fk>g?A(C[  sWg*tGy0OK9վKގ?xnfwϽc3'||}}?b'hr? /Fa"˻KF m,K`yp޿ dV$X&*+;2 ܤp?awYBz<M@haZ$㲓_x2hF ?S7K$~TZ~ ߝh`E=hol?ݜZg/7,1V9}x_%8CaмGaz|N/+fϞ>;fN kLν>b@‡?e/gy8-(+^beٓj$V@)|Ah_\/y;65\v6L+PzRh Ur\?|o8;[_{Di'I|6Qɵʕ06 Ѽ}qˢb6["5}ֈ_x 7 y<0.N?2.5ϞdEֆoӂW &gbZi?&Гlk?7M Q"[k2 K]&{JRAݷ ;/]8MѠN~$VBAJ^劘g_7tr}D7qBU福謌 '}d/|?壙nҟE#6(ɥ*ujO@ Q uX66% J b) ۱_W3iJӿfa)x|ژI{}lX67ӆƓɲ{b[7g6 W,Pf秺3h޺DŽ{E&(Hف9f\XvqkiPd*NDal/x7EPfqϩx#M$:k8!%Tb L5x'2 ʹ GŖ%Nj~&P)H(ӂ b 6Cw %t0ߪ&FIzqImEpjk3#W޸pPJI7xx\qZϵHD}ǼJ0~\#Q]f؉r"S@B9csm_&uh/ņk& JZPN=Ԋ,,)TXJ+Vx()!q- c쌳JPd\jn#a 4>M)p,tseZf8zGm*)ڲp^q[Χs{Zǔ!O7H 3BL0+!A,!6e'db|B^?6pH)$IE$s2J(J3K`@@“H&n>nqǒqk_?mba]%EpNaf sEqϹEP:nFmeYf6ca4yv}!՜3‘cijPP(AZd#(a RQIeU%#Ad#J@UR1ZE E19u+"%k,KM\8{Y'>gv@$_r); Ӭt&!I,~)5JPhk/Y^ 7%K.y\EEl?P$Y~Ta`ǁdA# JA)nE8ӝb 0x:k^@0! H ˥CRɕuM\ wrxXz-Ŗ F)fcӟ,V ViĢ^owz<]k-)Fziǭ\Z\Â'6 ^q[H\p$LgfTΪɼK3ջfӣ‹e21nṈAZY+d FVGf8_A>eI#1~Hǫ!Y,^HQQA3Yb1f?nrJGm&jԕsZ/8H&HH hRsd ;8U6˝JwZ}7Ǜ[JS4ȗgC_y:Rڂ z$pV觭(1gIZ [U_3UbX}_z޳߼}&{^//RU$0rg݉N@C 0^34Ule|q%ms+Xc_BH L h_8Z,t>'{u!'?'QeH [-R#KQ+{*}HocwN\RznA´GG,]cCT[i pn ADEogD=iV996mTtӧ;[3u¤jU0;RQ) h^ks~8m:bR9c6([8SYO8PØyfU\*iZŸd_?{q~Ix~T9gm.bog/~JHC|=3z!j$^.Y {X:{h߃L%䘐?GWw3F~pe.Q5yoGyqMU|s-z9:O` RKP::_&B:'kA/o()1U|_L%x1מ/f@*4B}rkΩ{X\ /pn'&^U'^ e6÷JQ _BqQr03f~"^{j9k; Mɍf'}ww4yHMzC7q0_) E;*YMЬT-/ ͺq}vl*kŕ9 Ur& |_8+,o T2O֬f ˴(ߧw=iz.;<]sha\9]6w|}s\pګ9Rz# c{ WW Ѷ):|fi{HnՑ)ʞv z3hy6,_6#56~_'Mѯ+?/zEWQ}3֗ WoqЭ4Ns:1tmk<+,2~#So\ 64a=یOU&tk'J̖x(]_rk%ga_m/`d{ 6ʀ8\V(M8=#FH%9&-=P/Xa\۔b6?$խWrusҘZ #dVs)?3Yp xNpfb^uѨ4]>$S@azS|)KBS$Q 5PPP]*J5d06LOv]/yC[ʫ<$cqA똵PK)tfRH5YTSnvo[ "U !9G8e=k 5%*I% AkdL؝vb!  wi~TS>Fo %HNT#Z)QsJyIP;JBTHc ]ZBA۳l>97nOjgܿM2}4Cqf =op!7nʿQNk[uVC "?&+?'ͭ _p~G-VP~L1kJhF ~5/ )ȋsəé< C)奆J¸zL>ޏ#~=Bϥb*2[T/Q{sI'41R%4`O&`V\9g=@&ғ$7ߖG\-M寓5=?M[s<__ԛƶ:~8L~()@rEZ4'w[ߣt(_~=X7ux"lMќOq~lVǞAd^sK^8 t153.0O<C9  mCF,>;ԂVkq)';'IARSRJX删((>3JNq-isP)c$pP8p6V;E!2.(CS0$tM֙8k6tkpe#Ά0K%.)8Z~Qz5obI|i=lm.x-7h/qљ=r#mY.2 eVu R[6"5{ciMGXEޏK16>Hֲݞ1&.gJFɃLWI,Y7Z3"ZQrgVX+*Yƛ[5>H'̊\;&BQp:* 1(" A3Ys8pXWQxn&O Jj\M FŹ-qjAo'Ri#\@&|#.IgbTR<ӆDVZ !Ix~Y&ڔd g71á?=5oLiVo?Z^?4ArXYOu448%şxqtFZ8G`C{gjO4Gn /Sqj*-rFw %+.X#rx$C[U arz$Z@J8j=!٘ l ~qh=7\i֣84aD~6:mazH+чxUl %Q‘=i>Ƣh)1dƵCM`:ƅulĵR- ?=-S#$0h njAVկ{"pK{?-}-_3rvg}"U>@)51iSP b1> >iL|i$KdɾN'm)nәNlT첸y^TM q:1pKOQOO{Rd0엾l:ʊJc.(]Զthlϐ\\ rճ1亄sNԣEܫ@@P%FnBbJAH!7Rg-ך@!&ji.*`wItOZ#E_Y]vRg⬶nC}^Bh7֫3sG%ErçBt+ǽuB ~!9nwa@#2PԔRXXLPH{iJQ%58(anQ6eQZ*S)xT`hoڠA))B2&ؾ~B'E![bVeοynsUPfպ18{LL`w2/Oi~l8-'{eBI,M4˽y0c^n$K%WIՆBLaږHr}yO`^f#fe;:6X\֤lIRWHNJ%j,QU 9H$lc( E Jȕcw*rj@R2;m9'bF\Ѧ1V-ESeBK9D63fN"ͣQEGlX\cGb *6?p]aDJCł'FwUE˫L: ]-!>Vt3Ar䴆"ڻ>.)19RJ\2@i"&C[vO_LjmN~0×lN/@!cj0iů~]-a|PRZC 'CD NI ?žEtشHVXX6?-Bkգv" y~^b-;i8p@r׭g]c=\D}Tt.))X 2imh_YoMNOaGfqrЎeN;2*G,&EHYq(m+M*{| tk l4:³$hЖAQO\xIrH) [ZŒޑ9#c?-vB995>X|ܐM|շUx{obu#6ataUVJhULU"n@bŋ9t}@NnC#{ƚ6mIBp9̱ S8lukp"˜//=9mt]=#؝I8J;]?a5Gb8Q5> U߭ R_ⅾlkr2@ J뒪7VU,bgx͜pM[O"C#爸<^'@$6TϱT6.XlTLBڔ *bD5""< 7.Y5Dmst8FmI-EԻ\7sGįE5K\\ZN)mc8]#''  * K΁z s:ĂA!P$\8x8xmts<A{_|G|eYZiߪeևeM\[fJzw|Mus@gr_$ ځ% +) Og3BRBBbB}6 0@6<} 7GA))TQfUrVɏД4\ >dd^=mώf3+ߧo-ߕ_I;Ip>}@k fgwæ 6$tiNvj0ٳۑ?^JVt)[I NJ [k>q|XLճ5qAs c)mQ8ng̾o=VIEAʁ"ڈBtZ\ :d.x 23f3msWg^,r5ƓoSޣUTcSrI*$dE:R+[ LS4Z&IUZM^[ rP(ĩb UަL=_ zanF,˦GडRjlLl0 N U4%SYuiYy_ưvc\=ʁVG(>(M$IF-3+Wel-0% g7؄|lYR[W*{߾ ޒjɠ9J@cWۼҀǪ`lNNMs<H:R0N< v:il%ҀV9-SМT%y)׹AI9Wkb竦<66n'0)b-jϕA'-;m.PJ͚dljW -I9DJB j]'2ۦBﶼQStE2[ [#a X|Ej* f{#F5bQ2%5ÿ. kBueEȠ9ZFVI37+H̉WJYGӻy(yA//ZUrZPx\e4=rs&k-0 Hm՜yTOU1> ӓYf;q[GȂa׃Kz:lWFDV؇NqTX&a\3[bdKB62zֱM <xOL_r]|1er&oSWrYNN~[n: wOc~l׋lJ|{ﰞ}>ʵߣg#/)h yvDݎ{ml7Ѯd;#WBZMQH!fp7:=k?Y*=uf6/: 5<1]]RHiom ٹc>v 􍆢]\R8g񭱧5I!7\[T$S&C:Ge.ZoKxP-"D[L PG;` 3ÃϧF>} ?3WbZbҩx궵]Ggz1|:v> `Zk1bՌ`Wi7vPAwHeҿ&mPS2 _LP CW08kl_銹-Ȫ5׽Z.l;lEX^=u<>_nyVKezl59Z$Ўkj( 064SmDe[ u+fjBZDT#(a墹B2P )XI)f~~g?^^^^ױެ-pj<)6".bf,AWfn?.Vo&_V=(v ͠sPL!D!'Ѣ(4Z44ZxTttF\0TU1fKgϘB*b%Y$ⴠ}+O3!gGgNި{n ; r^_{PǻcLfb5&x%a,/q]` c2}3W«%d-/yFiPNϗ=V>Tse Ql 8=a^xyi,|yMP9@ `T BPޱnm@RRnt-NAa2,X>¥5^jW˧O&1ST #Zw3!kl` ATI=W ߍ~*e4^1-FoXb][s+SzxGª<ĉϮr8ٗ؛kdHJpH.Pr"@n gXH[LBF̀M%d b2#B"3RjD$9D*WK%D)26gg^ Ԭ:AR"8;)*lsDt ێ$C sʐS="4I 4-vkK(!'A)g2c-7q`B=ujpqZv6O> :8^Rg9~H"8̳\P$\Y#!Τ6䢕AvqqYc2㲥 \gӜ U]N̍%8 -Td BFbǬv9ث8ӶNDÏzkLDU>ٰ-PfUwq[Me3Exor ڌhEy6Kc,hGcvC!4l βJ[AhM}O&lX)HP4D4X TH@QWH "P&K[[1fY:y贓,KsI(r;ý&.#2%?tQ,x40dW3њ!w䬫y'b|JAd#Wlx P2gRZ_M'5`J3 *$TʼiFIFDo^sיRR#'vjR XiBJJFӞa2g{Ӕ`1y86Gѳ},YNuI#]qT8*vRR%n6;KIn 4|1}&2-u8~t=,%9TƗOuiJ>ßVc\`k(c#89 h*R)xDE5\wۈsݙ=+GS4 J=֥B22iaiHd#]ic-HZW3pПp2 $Ȇ&H<(p|'Pz{iUG\`M~&K_~Ĺos|^B@j৽?WЄ"t˛_]btJ<姘Yu4$\x7L΁檀o Io=TnOf\t3R1ͅ rtBo_ո\M񏯊˩5޺$OpIJ(,n,pSq>yu-RxHj?zɭ2G( 0+w`Y ߼^cˀU۔l{5 [Gq0D88Ѡo?W>2UsIŌ~uyku//έ%ߨ]&"R„WSgōl4fP/MWkYgt룥"]P#wIf@Zn#_z3{ !y4))b!A*_ko܆|L/~Ư/yޢsjy|g6.]sEvy gLlƋ=C(}5J6]q}l8}m7҅' vˆ{dkCleG:6=Me:5-~|^Es .6fl*D$1ж@أy&.{8#MN܃wa<cT<YAɘL4r*?m#5 {G@ b@F/J-:=wNҫꤷ bݖs78)H4H3'e$^ ܃& 0#Nf\~3R&f(jW2{A2{%v1lGDWUt,th)amᎮ*BW-m򎮎`{7+*8E4 p+gE-+@i;]JEGWGHWL(Icr"2 pE4th5EmugWHW\+]`!U4td4G\Ru PʈЕBGuv%DGCW.U@+Zd0T#+`(*%tPrҕR ]`Y4tp ƺ hl;]vtutEUS"=OT*L(bE:t,k.!BW ON{R0 Fbş7zZ-cNh&9J)Yʌ6.OCɝ%ɓ]T*g1S-G tpq9"u}fZ $ L)Sifԛ4'K DlXp3 f26C)ZdVo`vA<"LECW`U@d P*;:B"Z1"+d<:|#%#+ڕ0#:\X JzJv1Gp]`AE4tpՂ8HwgWHW#c?]\m P*::B -f, p9jX NW%ήKQ|S :a|l^lKT|w\'!iG16?6,{T,3*Ն;kW(FKw\ L2h,yFK*04+IJhLy~< 1izwIA%oy1+i©BFpr0OD0@Mq*07s/iz S ?>'09;'7 _GqO^NLz^/^rrd9S*I,-h r/q5(Ai<6B`Af>dXC)[T TI7NZ/; ^\.'of-iY(wUJww>]))jm"9ف8UC<(@}cVdT!P:Yd}'.]+cI0 sLGRDeJBoCl2xA/iV @++W^ՃcĻ~ mskzõW¸uzTa%ȕe$Q+yь;J`yw}?Ar3aZAp3]Q>f $jkbUPq#$bUfUWri*Uag 70Z 1R%Jg۳"JnwX-sʵ8lQ)| g8l hanRvJF +ur ]Rƺ#+J,"h*XWSNWR!>f9 : hww (yGWHW<0, pZz*lYߎ#J0P- C.gU@{{Ԫ#+{?!fRFCWt(U@XGWGHWTpDDtLBWTZo]Jܲ]= ]11U㱮.A: hl;]wtut%Q92c7wuq$ ?.ʬ/ `P'(R&)(IW>&$HLtTUVdduEV hC+Aɪ"] ѕ7{\YjtJP\UL-QoK[+A~]PtJ~s}CW7n9:]erN UNr~@9qeL+ r 6 4;ͰaiA[?YZPf]aiw{ mKjݹr4wfe<(RѩTҡ7DWMZDNW+]=EBY ow+h+t%hc:t))]=A҆ ȕJZg'HW^mh3t%p ]mtv%(9*]=AMxCkW7mf hC+Ayh6uoBWhÆJ7ǭ&ePzt1 v ] Z%+]=Eʖm1lnL1RtIU ΅{># < R2ֻE'=wR= 0}w{*=q(]4MJ_>{ Lk?5%RA?tQIwvp:֦͆d٘'pmފ] JO*s7DWo.VJr*]=@7;Gݙ铿ϯ_u`G1Gc~<ٙLWҳiUn/n`iռب5Jlz,6!;3M&(;ˤs3cXc@P"XXg!8dGh/âd^D`֙/+ޅ *9dS cEn1x[ **(:;-1y9(L4eől( J =TbՇt%N@2h +6±64CK`1wuU6Ņĕlc#џזx^*&w]+&l@6}j~C*5D@K-'w/CAqUT,%^bNqMުro6}'CXexˆ ̆ e@lB‘l`AVࣔ,*(=!cT4~j*3ԩC̈́ "XK 2ѺFPB]܁vFnTzCݕF3x@LAAO@vM!@YPѾFil}sي (q4vN֢'d̿ s Lp^_]cY<.NeC>~}W ުI7zi]&1Yh2X 3 b|PT8xiLdM.%Q!Z }B ֥̐MbXtLGɓ]@N`ǨVZQ^ 4I6iy],CSu<F2]^c.z@ }Hd Q& q'2Cp/},P -w v<&{kl4FNG/u>}t//[y'0TM4`UyĖ,OdPY !A]Xjm!/VsP6ٶ债 +& 8JO5Ҙ.yAU ]5K qr,-ChWjюaQ9!'ZR(5bPK`YD[~U Lj9jqVyBr΄@HDe#{ q#k"8Ϊ$1 [ӌ,rd@CuqB&"]G 7åUeXJNS NJp"*iBPʉ;qPk|?VVHWYtMSLD,Y+"(m@;O[7W3Hs,efUI e ?8%% C%/. +Ϸ@ X.^^[ R/[; }DOH7?,،ы6rD. }nPŴf1hҜ\k-)[7[Aq^,R, j31WrrVˣ3 ՠ\%>/ŭaCxJo?a[mXf^I`cfPw}Κ̮vt]ȷlW}N źIfyw68ߖA+Z>-5>d]+D,ta:rNJyZEL?r|r% "@%ӊ',hjɮ(P"XrF9 =1E> ]YC/!zd*<ۖp \U= e<\}sUK)#sEV1WCPjzJ)\mǥ檢w vj0Wo\iSvaWu#?}1W檢j0Wo\2Z 'vEpc_UE+MUE`ޠOy ~hUrIsUk> `u sUњγr+#y0Wo\+vEOU ;2XQae-+Rc0 W&M5檢3&\%J0a΄{Uw'2 F~ZǕη'@?}i-]ܼׯ$Gv|@ s >OxƩmvw%lW{sq`6lĦ>A2* R xEl|p ̨8Fkl~stE~ހjԱq}+ur7ygk+|%վ߽wy z_a*o//rMzlNy}8;}9]6ӃLxȔ#>WDaz|#M>K㫐vĸoAftq6_Kh[TJ&@jˎQh&B/xۻU3~es2BB_k{ F|gk=[m#}o#gO%R׻!wHWOWAV/OO8U'>y]ۈK "ϣ˳IpqUq6 >_lOfd V31zGн6vgs{]6KkOeϞdi΃oKY!W֐E*>s[]kr–3:9iJ ! ɝB˵%ZESrے\;.AH B;0ra'L{`cO}ap݉sBkDE CٵdZ8 dgѰuF][f=yƖ<9[K#0A)k"Ys/r`ZDdB{SL9^|_hხ'8tgYYXN9fcsZ.uDH%8S# EN庐&tB+}o.h'b9H8j2!hy58;q|;6(e`;GjX"4 ,WY]fi4{|Qp+Uyý|*D C$xK9 NrH N6DRHJ y뭰>Vuk}mV<)miԴ Ho=z IYˆd([֕<1@-4;"RN&a"q\,<":TpoA%#2T֚Beth MzleB1UJbT6fbBd@YυB$%BdLRc5|ȵM 9f` bIAVZr"_+2rYSz:yɳį~T)My p!ɝ_>iEmq^ݡ+t$@j\;Q` ъ=bۋcsO)6l|I-38Ngo#Ho=g\{uѠM~+>D:$te.lzfcm[a3՞zosCTB,&fYf-wȃx ?.}7ͣ{xVj~&uSm@T=`$9$KNKX2@ϐmN ؊uW2[ex8=!k9]`=*]o9rחt]/]Q[,-䂷#ڦRH1[-L#(`N|sECEоj:рP&<$)sk|6^{Rń*p6hZEx0ڔ`0+]NJ+vB:hom/#q.|\+h~B'U :T;U3_~P+}.ڐo1,AF4iq y84=e{5uH@ v%&`OD*`&bEJ/ReeBFVkT+ތ?izRNXW/B!_V$ I!DoC՚1c& ɸ8pb8:ߚLuMgOHN\5%g\` $B c.R!}AڋY>1yW5yٖob= 3O&'DI!ff#FH"A:P cMX^hMBزv[ ƻnz\ZHLqɂ!"dZĜ"KIa(x:-d)OtDSEKАXPfl{o\U3Di(՝A9)4ghJgI{8+yɮG`X'ψx BW!+VARǐ45%3==UUuD(\!r -%& J=DZ1m厼@}v pT`lJDPR2 {oq4ה`19."٩ Kp3"ا L%ۋ.DpCLD XYhLtl, 2T l2-;}K}8VaVa)d̙"6Q°~ b)ZGCYBQך]Ϯ*9K]0Du鋑 ӯZ10 bD/̀oG7)#=q+*5¡(lV90zO7Cn6 4&+ê#7O._k>>^}

Y:ݼKG3XwD2@?4׾nj&*GJ=,]s'3.)wjNYf)kfK Ԯ?+eR}?RZPĒPXd \ )yJ1Nβ^ڴ6Śl&!IEUrj{C4$>VPxJpdʳ%5iz)ͳ/&Vl.Ex`Q~]/=4y,ﳒ$fE"f8~3g]85~5 /M5R "diWLG17R^~zg7Yn<-gYlFY#moʆ5񩬛-{VJn_b|.ׯfvy{G*Ř1TdHil]:TuPzt2< 0jp)Ng~ A3w# ws6 WG\wo~xzdqz%LhL QJ>Dg} c=9T'Ok .٤B^6;½(nzSWlo[~W<_[# |^'եg<:?.zfE5wqy&âb'0@7Y&?bev᥹Qؘ]f/d$?9Mm0950f5Ƒ(0DΈr 1@] 5CS6Zv9z7W.2ł, |BHۉ]@'] z٭ $ )BAsrKqSAi"3 ~5sOmrײ5~P4d;eǫ>GA:BΥ#?,V$_ Dp\s=4FDDŃlG{W}1ُ,`!.D䨔x{bRjIh3!"1)[EJpn3<`.sFQup.%˃x%#^nK%Q $2hwtQ3#6-"HQN Â; m$ )[X1c2b/Dk45[!-Ӆ+|Z*%|p1'̪T(IAO`VD UhÙܰg{C7_eR@-dиN.2T}VLk6> }} ͩU N0tu6αt sBșT!W޻0<>Y:烠-~aƎpMBX`bSEœOOr~M>=~+E3އ(jK32ܮH+\zIj=|K0z %U/Gj e[jRTu %N4fN/"D~yL5RNDTuR W:1<9UF-nߠ5eT0e]>(rH1gYvDKt- WVCMcԚ SyF~\*Iu6xӺw ۬]nW-(r5>LF5wח#~킭W薒ҬWQ~tSW6pcֶ;$۲`iB6РڱJ4rx9}8ά7II5 Y`vz=/F#O[cIQI>oD,-.JN6:>&_ ֵWH+p3ٝ2:^e1 k4clm N^XCA߂6 }suرŎlgF׈@sA]cqĞ:m=w. y[ggO]_V=4tcE AaeڿWgRX_}9i>laHQ90jonnauVt:xͧMG,RFmk:OvT0t4 ]Mi7oE`:>p}絷$o3 f_7e<օv.4ŅOfso82~ [@>-$_Twn:XbځCo(4TheT6`]t}.*F;]՝Qi5MCKZ0PNa D6\x_87v6$Wxaֆb~snqtߨ:t]lmP`\8cU?[uC޺߫7N]E䚮W7S󱃊TKfO{pñS11#}<ֈnFtňň\o[(P7ƹׄ~{1yjA55uxjj0ʐ 1ƽ]3gz4`{l[IZӴ>io87=UuC^\}\XZcE_XZ5U{-jinsCH H‹po>=KƟ@a{\-bw[ljV?"8kfĤ&\Lk Li}I)<Ԁ:I"`rtŸi]WLEWsE*ZTt/"tŸA6f]QEWsԕm:[#FW]1m(VEW3U~㯻_U 8CX}hY'?/VoN//oDžm#/ّzUaUj\8^jNlZG-ߜwt4T]E4룁=^QZѿ[A"bB7J CEXBFaQmӳ%\bzLk}bʍтj!ҕZ FW ZuŔ!]PWI"`Q)bZosS)FA"]13vŴ.)},ut^1AډRo̶]GWk ]13θݩ6uEPu]0j913L )] QX;9C]1mQ+͕jc*8D [;hѡj?}+hڣ n?ﯯ$MuIlj6$zmB/HW'HE-EWDk]WLil uEtE^ɉWLbJW9ʂ+q=JF*AXN}emPGWLM u h銀3vŸr+j%IUj%IW (FWkU+6w]1EW3UpGAbɓ 'bĞtŴrSEƮZ'FWA6v틮fM'OO* ]1m>"ʠJt5K]a%wva *m-66Sg$\(1ygp4c+YDgIueY!7RKp6BNjv7qȑU 2U}銁+Ɲи4ک:I.sEW䠤芀G EWD SLmgAteEtE]1ShAҸ B1TLH J6f].Yʡ6 9"\PvS5(}ut]9F1cWDk]WL銮樫vFB+ :w]1% I%;#3ȸ`izQbʢY*F7f:([j+Ih1L%Q١q1!̋%{lF'HWplԾtŸS”FXt5C]YeUSI(1b\+ &w]1etEW3+֓/|K= ;i:A<]QWN+ ]pV9)bڀ)c(1+vN1ӆ+8gj> ڇ銀DW+gi+t+@8ʙ$\7u4Z}tŔDWsU4 ;A+Z+jnl'Z(6> ʨmԱ!ϒX2򭂽`H]:]^掴J6PHL“(㽞|v7x3.pv$vЉ.٨#kK׼n +}Z٢X?A< 8x9"HѢ )3-:nr+A7QcbJf+1(HW p()bZuŔ>]PW]1p@1b\9jʃ+ubtŴP;QF[9*8v6Z׊Ӻ+ ejBg$u 8Z#FW AuŔ QWy'3H1ҔeUx# 's*(j[[w$&x|/,Z!fj֩9*,)qQ$8qxĽ4Zh#24|F'z#HWӂ]WLB ue3J +µSoL:w]1en{̋+o銁A׉]WLXt5C]A ]pD9AF)+wDWC+节%3ȴ6+j!j#)b`/gfqQKF]WDi]PW5$]1p]n4btŴ6+_t5C]!:3}:DOIµ*J]WL ejQ#i슁 3nD)"Z4GWLۑEWҕ1#2 N hs ` (pn2[!KI +3ϓeLcJ:)Rjp( %zuǺ:TmLc& F? ClWЙ;iW׾UߥiRMsGɾ}샻b摍9[. XvILQ_ҧcjI9uq|4\=aT (!=6!%}lkׂtMP*%EWLB ue,]NIb`/'qQv4ʨf+kHw]1dZ)sUtu]yc2X]btEQbuJAzgM~pIb}ճt X=ߍ)fE-zKJTU] {wGP6= CM{g}K_1㯋u_YM`,⿽?-5p<_,һrуux~č(OX꛳[ovxUzTԺcПw֗x۾jv+ƓiȮݱcdx\Ɣ߮jqrW/W/[޻IƇt<:-ς9`tw/xa^G\of5~szN||R峍XB}§%GZ]ݜ~B{ۿgy{^π> 5§w/Ο)c7ʃ=Z?|B~#]vj-)[9%2T'Xn+z @mtpʫhޠ6_ms%Ky7"g B>*t{5'Mmi ݆3[ Ү~kˋud΋}{ZGu:|x~w_Y?g7iFkd߮3\;E|VU(bk*{ }ql/[F |CAdI[㼾LQYWyTxW겾~]}sBkW#Uݶ ^%sپۮ}މeusQvY8[Sܒ_T*Xgߪt4j_JJA=Co D"nSD[H) @NQQq3ٓ9f|:1BA|^̈$?bx|?GḴrcqi+7{$\ט%J+4auEezwAV %bxj|\Np_"[WuƳ'7X uA謾feySfQs_c'oŻ,"pڑqq19DLZInױXD^ǷHڹ`r>y8eXJXeuwuIgj*QD?Ҝ0n#S64Gэs» ͏KV/u5O??O+:'&cQtF= 5gb$'r()lNG\dGbm}oobzLIީ i[LtQ87Nw 7>xX8$SwA߂4D#ZSf k.d8)1#db_*ln eވcuRIQ`qǪ0Ffq,q1qNƬ^pڿ=ԱH@hYA^Ƭ̱J:#10{ %aOzz&̱٪}y}fEb51F6^(P\X%Lj] e)]caXM 8bhƏ*p90Z&{g$FpKf8Ƭ^: :*1fӐp0ęc ˢ'$KzIJEqc`29vuN6 wC;O0tqu6ݧ\P]pQL_098PQ&s %UaNs2;+ȁOɭj*eBmbQqY sko9yߍUKdrN=@ef ~Pe)++R4O˪ DJ%<<^n5g7غsLbz<¸fH#x_,LTbR)QoYAxV` RTI%)D] Qg0+Dê8Oh<ϪȖU.:qAm[uNj]fO[P.>Gi̱@Ru<_/v*3g<8K`4R@ IVkd3WmiXM󇄉PDĀ ǫfT 0ՏY/1oMXAy0ۇ~tp26k4!&+cfR[65R36@jn2 ۱ٿ<?ڌV?n a0R~Z軴wCQbiDR'Nh=QHbU"VHW"%GQH^)* ʬ*˵c'|1ux 9#[Sz/h/P3>sNBcVz51x(ا玢ar6wl`( d9!:\_y}HAsǦ51j`R'*B~`c7n!(CA~t`:r5a:LdtAB~>a֞ d֢ʸ ⬌g:r恩paPc%.} U<Hw#%'K1P1?wgŲT׻,/ߙŪ;vws.fq.|{p}L $)"l8.7N͙~&e'J`0B8*N~t! {*,(" )N:-YdVReD)CX "˄RJ[:FHu:ðELym9l^x61xk!5v,Z=@̈:F׃~x~ݏ7MΫˤ`DJ vL'V&ª9o|(g%lG@ hlRՠ>\'7xM 6f@>9/eM~輍y ASq)Vì˶BԛVݷoX5XYdB1k]\4aH4rv0sߞJ YGi]v[In]B6xe-<zĕh+<;q: )U91(:MjIA5α\n]xE\7tJl|:3Aee憥|ah>»W-`ry>GٿsOLde-ϦՅ EV'ퟆDfƠ? \VU%|nZvт%'LnRR @tE(]}j3#ܵ9}jmF7ЂMBc߰ߊ~Ho.&(D^ rՐjԵjw<7~.qs% *d+$'߄Vl-j{hpNX0E] +:ht(0<}^\ggIq}@0 iI] 0Ck[DMcξԊՈ_Ƴ5Fofs#*WxGֿJ6\IA4fm%_[װGZtx=0d/Z J9~DŸ*SBT|~c+`rN=gƐGwϦw !X̦U00C4L)Yp;JxK7Zͼ>[ &GiO7iZ2ϮmFZ[vß?%2kmvxy ξ, Q'㲝K/8bK6e^rn*vJs-˰H-lP@K?$ !Ҧ29v:!t4ʰm @mN\pׂ%1EwkS1M9!q[G۳$%qhB3x.>'Wu`JZJht)?ٱ9S-&ϿlժI_?L RUu .@ $犈֠5zept%no`"L[?Ŭ`0FH #]aP@[@, XAGqUӋ:PSn4?<* EՀx%JI6W-)x;G, $j;]@j cE<^/~ խ/6Tч-ذUCM8_"ek$<2$թ@&s'>Qv1ߗP fcPJzmWhԬw Cj&2A[.h],Q>ţ -@v -M̾5zJ G*Zo&U#d91 &Aܒ ʀ|;?hs;GALăx*XlU˪} "IKK@pxFjJmAS]<<+?M|=b49ySڹXiNٔVu8]dxy3} ;Ҍd,cg  JxQ7:R4رڤ\}s$t-9Ak|en`F h̝O>9ĂATO:x'[c@e2}/<&e^{i-$AQ S GrjCSD$bM.lPHi8/5lR(5!³N 4`w0\SDcp'> ss(V:Egha}&|izk)AAqk|"0\\u6n"Zɝ$VeH yu#8rs d֢ͭK]#6Mlcլz9\\ZS2˟R kڼǏxK~]9UcOI3żX)4MRH4,dQUVX<,ŽIcd_940H*v U+rH;B_0%g&5dyh%XA z C,b%Ȃ#k}8gNPs5h*x1;؜],ӹX{}zk-u19JT'|x.wcvmPY<, }!vG//uHs e"iwbs#4]5SD cčHB C-Tbӷpeck"g񘃎q,BɌ`T(n4pVd0=/w~xqC^B4*w7*W(^+ IQwNMnCޕc$g7l%79YԀz%GU*ĝ]/ T9.df%mScNx | |j4D\{~Ve\ӻLh$T /cdFP0we'T{15V@P鍿eL|l]WtݖU[>WًG9HwyFV7/|uӳyFN3Y \rn`=r vABy NN\%Eʾ􅹗;1:%@(IStQ|%,Y2tT&ݎ:V7!瞧y"}޹ \Qj3}̒.uFɳ.7㲠bT_|jܗ}K' dWűzNͯgfy{K|Bg+Grx7KxFd*䔺EdgwJ2g$m0yIqDdB@ZF=?cOO{߷QM pw ZH@8lHbclY/~>u㷷ՠ='Dŵ(I _!޸ߍ.Okݳ 8M3ylu7Ӈ;{8ug Gn໥va睝'̿Qe He#9ZK~R(]q{8oP2DJOX Pi:K)8J1ÑܳEgs t.L Dx2IK#¡08q: D>hgTvHXBON4,'YY}&F0}2ɜ͔cJ(TJ#QP Sܳ;Ly6}(i5)'s_[H,tUϽ)M>bYMN Hr sq&kՇ@յ/D»E>rb5I:"0ʉq%.9Uc8J"jyw%5vowVkwOcCvl qrd2E9BeqEGA.>&:T;"m2=oc:>[7 tjAw/fKHW*ak ^h׍+.aYncuV+| R:cP}g^eo+EtbBH*qsdISA5w) U3VIanK=M"!C:utsjl]qt$tNT.o_%D:MݤƸیDF7lK D7NM$  XӇV+JAga4(Z۝ MUB /nADM;4ӱ wv|޺S YH_xHNLܑh9BX1HLg4 (UB_yA;iw>\o $($260V"R@(Ͷ7+lZGލB/4eRR#S/9|9S9ٶU$9κw wM؎er-g,wY]5ZKЀV[XϺgbf"!L9:?6n!K2"q찆h d-KfHSK܁+4.: Nsn4H?:%0a)F%tz^4pBk^VAr|rP~HH\m^c1\v@JD\W(] 0JֺtvdHTҶ1DoN i"Ȥq֖j^W 'K*^xWDpZ%Ah#@E,G+};T lМsw\ ŷ@d/-z)a2=XThoƝb;v48- |ۚŔ{R6םw˧ /NcIӛj^ {4繱v|x=!s ASu)inX!B[$}񥁑h+ҧݝeVeL8UESǿ}@)ʒeI˶~i1iGO."q&$f8y8]=.@,), v\4}UI՛RhWb}YR树.AbL" RUJMPZ-(}:5RP^PSd (h ^yn;ȋDM/l(4;N23[Bh.=kh;Dl1m'28=Q+5Fr2"cC ^ jhwC'c6SX`V7VlR Kq?M//?~7*Ep,5DPJA 3UVԖ6Wgw`Zz, e/UVnk=&a1DiI-C)H"gS4/P =elYl;-bNSeMPUUeFJV58!eqrA{Tei~0bL)[(e>m!WDZ))3qDQBĢ\o?\k.@pOc o7WG :xr55L|Wd⣄?=\T`MŘBcNfa4q؆'||{ETt .SZ'Z >i壽r;w3[#:é/=|p @TRvVfX- ȩkj;oxGhUUۼXU>+CF2ځrsڤIw<<t2Kp?mtU6uMi TPQpkQSDzbxR6 LvHƻzj]y݈t<݌(#b̤" 9ޕ6ci1p_[ T`(J\ql,UCN"/Ŷ BRDYc5H`i"QkT`c#;5ߖ ``rʖ8%ؕ4x#͝EHV)gl]wWC@D2q=fC\I[XhgU es>-EÁ lt:,9'DQ yWI<fiƘ!' B;8F_B& ɧx%z SɘKZy"Ĉ;QwЎD7Lb'ͦVy$9e Is&)ѵĎ|V"óFyf y>Q.;뱀g?#?“YAܸkpٓ s-12D8 smfQ.814i9ivi*m·7-r5{Luv&B!έGrI ^DЭ f~a%7})V6ִ!ݕsBDUN&pl!pQMT]"O\vGO:UsQEWARq}&9 ݮE4?R P&JI{#Fsv.ʌᲽc*Vy_;j9jN)Dx,DK-$Qύ{}5sboRcIFT9F$#N0q>)]IT*Nl.R"M_jo9f"P7tSDbv]NFi8p%Dn-1Zk& F~BiP$t l\ˍ޵BcW{:w.{4h>~ ;l+5OL(bѾF[r [Gp>q^|;EsR8r΍a5ᛍa/-( 9k~I8ÎGe"2=J| iypAƆKUV m@az,!}# D(wISȤ4>hn{Y??h87:l7P!)LK B%qA6?[o|:pRXJC2.ujh˅Tģגrȇ>W87Ľvk!}*ap(VL.V)Ee0fР@A w!U _xÕV'>dlred<®{`5ǴrXbM[̲?M҃1[wLʏڱfNݫ~)#3LuY O|2.oaϾ{}U.t}":| Lʋ̚"s/_7&8s|B؃f'7tt3NC퉌容p\s3^(cKJ>V{u8w啢i~kLHU7f}Kjr`vz2ݧGR9qʌL*0nZr2#5mՆND{1f4T_[?f \BC 56uFU?B&^igb/lfwd_<6[u[E`:gw/nZgL_U@QB$OGgߏQiB?'< / ffG>-^l{rL#qK{3 7ovOʜE.í\Wz܆R˛y|*H3]ޓAs.ܡOU'9ưRr`HfX*UTZvn8'x%zof{5ٯ4S+031z#Fd6M9"4$bÁYI'^rnVqwbU=uSKE:b$k+s Icm{20h~&{=Ml֧ '5HO"#A[a]V6 |A&REz|C xv}-ZX}@iu4M^ XLhSQxpu:'u:JlIa<;FFJLj&iJbF޽Sλz\={%Tl`b̡p, eWw+lF\x`9&G,?7Tŕ?r@K!tjvN-D'*`^U@JH?tB$!!%b:xQJq$'o<-ph̢[CN!:rqLqY}#|\Wݕ#͎q$ex&5#e@Hȁ<˅8T;r4Jєi- w ܠWwee12f"8$[Qq^L= 5_Fs0tñ Qpnp_hU* 1P)WQp=LO֜ce^c|!U) NBVU (uUbpQ߲5;)Ӑ Ѥ/놂`l12M~ ?Dq6lWMA;Nyg6,NxaSAJB$ l\71&y QH m.g]w^au^#D@xD.eκr=.hc@Kr@g=rVtb TlV`g*~%l!vh rHG!yKqTF㑃i--!耱hcDa3d`6[p}J![RkX1;JwWl5$=:{+67_^RK9qaߟ^"nz G/0NWp?KܮטeXc_<Ⱥ+'㲌0M#բ1j*>"e9Sb+r\ ҽ;[V hPg6K EF2jKo~ ÐFv6u]ѴR &ݧ l\Aꈜn*^r;՛nSC9H#lbJ]>OX"F.N+-s8,\q$R-BLN} WUɐ<,c^LGtG햇IQEO3W1\7]*pa2r߳,q"͠&|WMXbvzn[N̂+Eoޢej]Bk]bg5o ^~ʌc^o :RԵԥ&J~|hROziNz705X\+Wa䷟#x<?߄HO1#Vc{p~VypGyiZx <6(A)4q&*I?lܵߖ/_fyimbX/ GZbK#Z83F bSL&{ p#~`G|fw!G4?OS;ahW$2P(lԊ`(hs,q%d#'$'(Fh/eqVZeQk$FQVWl6'K$W&ň1%1gI{;+lv~R߇Ax&Y`a)qHO5GJ 8:8>im"~|JdAP𴴌^*X#"|mR/,"c0-B.>zcİX;kErh&ݵHȍik=6ZĨTjU'gd `8L.S~.vk"弴d.Ph%|0X4417?('5S8ʷ @Lݐp(@X9Do'-`,Ix񎯽I2'!EkO 4k3mMw2V i˿d6bιF8m^ۆDJЫ'f ɴoET,_H¬;'fN-u4Ұ4)Q^ea03bk:%q\f5 %[<_Ye8^b\hTa,'#Tz=wKUB\u}VX7\'I 7-"7mM=9 ,%@ bfov2ywFgQ8G"Jdz`g=0az9; 9)90C9ÞG? f9꯶Ԙ/b9@c~ }?rF{`ĭgyڙ$ಽ즣s스P-RV#-En6o^0sċ@瑒ScB4γ*ӎH|v cZp2B ['sECVE=0 QR-jrr06~] Bk7eSG|mƛ| eiq>NY?v~Hə4OG,5]TsíIR3$=%Edh%jytT`tdRlML2D * IXcA9S^Z,[L N3X0U3f@= mH.lI1~9^ŲDs{50[?R !7{T8TUN~+LCNG=Rkl\YRzVG_US $RBcED -MӰ EJYʝ;oSEevo踶)[ipk*pa$iz7#20̒bPDDĚQx֚&浄T'ZBO&wO]ۼlkP(i˔ B~!D "<.$"͔),7~W F\w֤4j/ JmȀL{@ԠL~0Uig \A4ZI$G8/\D-*wk}v* Y Y5fm̍\`wzE (6< oQT*=0 -ွ@yvR63r`>[ݞfv3sm~v&f˲j#O@t8 qƎBmn/߬}i.`D$C-W^F%jxd؏+wx_y.D{,jqn0 Ҩp^Ӥ&âexYW}KlW}*GsP5t 趍x$rBuX|fh{`|4 !gmtrE=Ԅ3x.6bqy!@piԦޚ\ ! `qyTB7?BjO-si 慎T !|$f)#@WŊ]jz3A(awgh*`\'qffe,)ꈌ՜RCgKϗ^5e˞}3MQF/Bi0@ԅijUw` 'IR*)x{ZL; ՓjQ@)޲]q7)-|;o.3OOƃSE)cӦk)>5%v[oD`d6\5#AnEv$Ӭ-xc:eTZHѸdRy7LTILNT 5%܋$txUe{C "@d&`%-|U\q'<7vIp:]g0vFpܝMwp/׷hr}uBMakM_\}W dd\\eknO-vCW]X2-/}kS{םiuP T3i2Cv0QJѮkFS")eCaJHBb9(0Rt_IT-gZR"UYs ^I(80Ce7$M[RFKo0[Kep!9~,x25~QeYǯTk*\ֱXSA,'06 -QcDT4H[!Rvh`DZU(ODr;[}fK/dX5fl9k/yCb*Lr.Rʚ4弎59wf /8c:xQvlߜ gX:X_~"Nȿn:z/~|qUzEOrTN"դ ~?lw}L\T{O:(C + LMSbT[.Lw~y0ݔ1"I8%CJA{8GzRxTap^l ϕLY%ֱ2GIٲ9+t bOv1& wiAD_!.A"4"/wsrIJΟ`:_.1םm= nI^:*h Ԥ̚mb[h{Rf1b7$bբۗ"&=-RQ~A7g)ݥK&/D/8X4vKTVwVp /" iP"J/PEm}T9k6$Ri*ts/Shqԑ? 6 O_ׇGݙ#wW$&_'8 T.W)t挺rۊ=#k,dI>c/6ŏuh6ƏfqeU-ys{#sV<=GUV+trqYq'W-c_6qYg0)tZpEFYf:XlO+>LZBs&pНSD(p奥9'$% ŔZ"UZ%q2j봬FA9A3#" $a(ݍ :61Pv` x6 LzЙzRiihӖ VԆ |`@bX0^ }o4#9 Ks Y4FRsW >y9AiinB+gO \xΕfi&`)*q:8q:& (HEVS_KkD2n  50#"C-yBL"p-Z -7QW(4gQJ10.JF2Xe(fY5%Rh jR͠n%^-*Ulis?2t'a߿xZ 5}#4' 9I$a` f&HVyDC Sd:|N]_f7`o!s^ky9uQ\4c5"vVZ-E gJmQbb $@3a/xrp;Af aϥW,k.>|9tLZƢc\7'n6E,+,AE(?|**k#]ᔖG0'(K@1\w`8v߷y?$c7P n?Ze ZrV8Ap9rP VtJ{IB@0f<`=Z2$| Q߸0}+]&75748\ E%Zvo:NLu^cIDP!V>=cˏ&UR1an {ON:w?1QRacSKF-5/uNjcz5거غLc ʈZk%9Ʀg'bR'MGSZx遶ǚg[]ާCCr~| %XKQp@S6x[Ffz:|VbTgn-v}ވW$F[9Ls\f+`r+\Nv.hͷi89Ikh{o)Sҷ<Y(iƪqpp=~aj@4gwɖs Gn^k RL |mXb P~hWrtœ7pdZphMhPQ<2eBuXlWCKi#-hEr)}Wv%a8" aGefÎcDvTSh "B!r!ƁD\\Fs ]QV5p^3OUO /Z BFN2i!s#+beY,ai s/8r$ KK]ڂ$>+?n?q\6zwhzZ W(sRΑmz9 Foi=P і Gڂ`]vH. s#z%UkKP-hKi%.ir ΒJ4zb9.&"AAKR9Ƃ6q rˎ Ю(}q_2+WZQGT27֒@!sp}j%11f:flm`88礐^4VaE!គ͇h(xiMUZQR`QO 5HSdѐHE[0 @z0:_XxpҲx{Ř֥ ;S3Nq8&t=궭j1)w"$,ĂP Bo%9!S~rIF""P5kugcnl b(jR>T)9687}7&&^CG;,OUZ` T= z˝mr'ekDHP8.>- 9=3w NQ  `WdB-%+Ҵm6[,=>a| !!-m\PڂѵH]x-6I 0D xŹ?_&|rz-lo'0kjn'[OqoFq#rwcW7[EEN 9⢚`1nX&kWFmP% 5(AL,|2|׼f&9<1Oi ml tt<S 8Cba0 BkRNG{#(͙C{=Dh=6>$Gcl>T&([ME #R-)ڍ 4-}dF.=Kx?~~a-jyp?~8V+:Ϸ2F:Xy \c8@&A)VJKHo;nwo*$*K@P$H$8}$bmdXrN;flDLXY:m]`6MGR$)ĥ8:&c85n2pNbBpb-/Cu<8 q[ni $W ýn,{svij}#C |*ȹJQpv)zZgoVX] 3e0XR)PIp16S;_E[3pK]VPZ /9HKZ1J &Z8&^C5nx`IR)cK[OtѮ{ ݄ /Q.^eZBPmІV۞-o`@2{>q-(F0-`+g-KPy;"&+c5`*$4˲ Af)~|t FP(sV*l8C `vxSK Fq$R믃8d~. *"%x@<uHhCI1 yH5)RpG^nX䕷hAA5.GER9)~$*UD%I/{5* }-)J܂NrFAN9[L/ .<:rFU!REhRa=9\1$uJW0{̄ OBrUY[!|; EyzE^t]TU#ZH3b+a8:ORD)qQHE'$jmH;WM6@/WDs/Sw8uq(:ٽ8M J2&ð\ʀ4æSI\}x6L_vh}:1aN>j^-_>HEye%86|WM1|]FdJgɁ-4Jך7Mٶ/z#JY-rvjJZ&s;{~A\- Di A Vu ߂֡Onc.( N&y ӧ1)q5'69ށ%K2^:y9$dS4#·=gby{r~4=y\ W҃tcu+Yu-b%@k%>흳)=ubEizg^䙦ՙ՜˂%.rL*dsfRx2D&96.$j݆笴j|of.L"F7IJrSswJ 0026~*U+Rj}6\v>Vv<,a:GyG&.y3v8ƅԹ..f捻.\poJVusH-ƊusnFߘwή/]1-o;iq3׃~ N>ȯN8gtšT (-Zt59o:@$KAJi]vDt׬0ɱG *BMTI*r@~MIi=2ITawJgH q{WT1\P(dU1LK+/jK)D~A0̡(i_bh-Ձ]K%'tt&K.)pHiARg/e]X;pABRK;ep@b2ݢKhs tanXމGIGft!֪ FQ^zDjJiqDn?,Ȁε gwE*rqu#+L î 14Q#YW1Ulo r UmXieMMqrVwm* )Qifv13ދpwui;]':9NmN&|DQ$"QdĨ!fo}y9>8#Sj'G(!=$Lm3Ss@^^<&'N͓># ^K N[˄L܏YnǼ8.d"kOqy˳.Znףu;wYCbxv7omw$ܵAm=}i2滶 WlVK%뱛ǖ^tr(pf.{TA ]* Q4# !/eV_Z'nmŢw|3Jķ%E P{m |,5t%/i?} BbeVƝ޾JD@Y,~t}i;v7XΦyoxJu54񆔫F %K &lZ@6|Q BɕjZT(َyl@REmJ3`Sd]_09C ?<~E&y9U½}a(Ub,B%d\X@QFPؗ"dd[j(.RƐ 2?(Q1dR0/cSX59X4*TW[#rgNgu0XIQ8[k1 W2Yya-=8q{=}їYkh6oR"( DYm mEc$rơb[kJ(UCB.fB-c:Hsи$ȂFhUl N5֢CF` sMD)9ɲJlJ8V8м hA @LRiB6P&X,xv &W4Y5Ik<w؋wGgLjߧgUgiXܪ:y֕%qlTYȋ1X͍yECѪey)7O.ǥ#)VH\3zgWq>0WWt%DbqBRXE*lCShm:j#O mP+Y>xfHR/{4e_FG{%d̈́dVzho1})ಽ,st.c]ٺ&._4MVaY=h_ɔ ^W.Ck3y ja8 M(V c~)x"u{+}nk.VsYˢ\ݚB*ŽxK _K,r[=s^b O`1M[9 ,Q((akLV)f4J֩ -iij/̖+|b|B' mkڲ ֭+^X:RR)ddTȔ#M6dNhxf~G$gU& 4Сb} -[Tћ$xJ Z"ʡ + 7 OVzj[+X7OѕCs0wzYakח7 R͘<[ᗷdUl*H6$|GTJj1+O ⟑W*J$db]b58O٪fWA?B: !]+c)mQFúF;ѧwMhO_R:k6JL̽8X#W-hxfoli[snGlerNnfGZ:si[:7eētu3={w2^M} K|(hFm[_/J(Kͷ%kQ/"w,ɫd˘lXhŷ?]-_ʄbf>  0;r-߷{:FГӓM7\j@ٟo}++w.lZ?7}1(kkɵ-%xCCH ^iɣѣ|+v'lelGHE>m6xvqNo8_kGTi@\wOU3SjEBbQUN+KK zq>tZ*v87Ģo1:|X#;)_t "ר}1-G߀Yx%틦Ye6r}@kV<5qZ7o wM@5X.|aI5l =7 :nQr9kK9'0!,|/fmhXcrY\dCKvrvՀ/yFjfpF `gcMM[^Ϣn&Fr'zqR 6SQ}*%{(=) b0ʷVvUXHfq5ʤLe*~DTxz39!Nf#UȖuڣN*eJrΧiczڎjxh- % "qϛ8;9 vw ~{1}҂,,q1߂~G{gm{d`FPo% ]2_2-uAgSR)pow#޵:-Ł~C9lRHߟ2w/Efe/X:s q9lb(!۸A>-R6F"T aómm@e,)Y6h+9fFfaTOg2c?ћɆHh+*il'i:{l}:=Hp 8)cP;;N`F]*SJZ5DQIAR(Pd"2au]A`-n}ԋ;>V9P0~)>'PC] 0U桅/ rZ'޼;-b^Jl]Z%tHrA`nѬ!0g//y#5}6Y3;_~C]&eC6qq+>wgo`v}[NF`EP>$9&׫ӆgu9Chs Nr0O,ݣ|oyeʂ%^½0Hνd5*:wb|1b"+a>wlҼ:vs1?dz<H{lU6q{j|酈g8%j_W%z]3X̹&k[Vꏈ>kRO(Ny˵uhO22q#coss_F'u)u*f;ͤUL:o@%f`~KoYd_xrAz ՆcRR;G:T~\T %JIF1JjKK ?"kf}P/߇?;w CbqZKu1hAT656zo0%VƐ klYPղfCCRdL9oheHngA {>|{q;w3γi>ǁ4z:VW;G`Nˬ"nH%ReA*d3q%öOvCJ8Pڤ1k3`X2|9pd,WU,P9,5,<| !b3ur!ZF[Ω/Nw܃`b7 s zߴ`U@F?*Ѩ:_k$ :Rui@9L` _9ϞU2?u}j@00 fkGGr&'2g o\3immAދ$4r[iPw?y)g&Y~6Fs}OtQz{jS˶RlQbh.2 Վ6M,>cvڨ۞Ӏ}-`\zx@+PlvcH RGss8<0%e.g-uy=>9y ^D|iӞ”Ľ-HMT,dsǥfBfJl$,&Ib ]SiE0fZj>R譞T]T^PeϨ9WWp@E_( Z|snfiA@Tp2YTFp<]QX4? :C.1I^>7*<e,y Rytw=ǫ{PǗ|x0O0֓KiIgdb_VoZi2IV5њZ'T%'/)G=I1$"sF=A.Uƃ3C~qp.dJh4Qrm)Ra5bV:m/7Y䓵X#g9u/8ѡP 6 I[TJ.s^SnhC:3$ hڝ/"4;9T_Ko9v|Fd WZu {I.~JlR9uSXcvTM+S5DCQ)4YR) LD0H,)U|%4AaKPupai$S i*hͦ! ݏ'R/.>xRbx8:y{ɫNiI\8Y$=g9|ނ[g[aپuI-brzelæ_wf]rzKSJRE|Gx{ɫx|vV K$|9ō~%ˡn[ۿaDulR(FYTtY[?ʟ'%ü8[Q>DP fU3$N(+;'\^:QlJKNэ>lf]G^v9է/:'b翊ukaqL/X\'ʑXܲ}gзSD@7luE toa^>H|Uhϙ>G&Rg=Raq{ɫ i4Ob Ԁ֝GQ#[6Ÿ#Eq~C*ll+M ':OظjqIsCHj7] wظ#6~>q";OX%lKREJ.pS  M'hN1Y 'u?1{~%6볭/D_*ᣝrTiLlٽ\ǼŏC;^v\:-?_/6 LZ}4k@E?W)yآG$r{ɫ()Dw iHzS-ʝ J׭o?, *< }wոlA$W57DȁdTvhREkhMJwg.m8d. hۉx{}UjC-Z5Fk[?Xƴ3 DkT݂͛G [h͛֜`dv{! "r;׼uLe/U;o]I.|Śgzl 'F,:8-SN'vs60oa ͧqo3`Ʒ".TV#6` > }1hG[\Tctha mnѾ'OTݠW(kyS.)y%H'ޠ3}/:YF?FbywLwtLuLXpNLQR X欈18TׅrwuYZ" ϳ$KwG2짃y'IVK y>ϞTݻnL];5&̌VL9Qл%p뇚<S2k#Mη<_#MYZNJTh~pY>bfN5ptNl.Jjds.갶Eѵm Zk94"Lm o|9&w 8̇O9JjTuA&lԈ8֖i)\,2F{IThHҰDAX-4ub77-}89j$û\)4ɪ&ZK#$B5 }oy&d?v'{38Ff/M}FhS}*Xl=j h_ 0H̝lVשTF v38vRx 3WT'' & yQAR {[[|~Ȉ-y.nK%`P}Ҵw&ZHPR[*xѾFzMCG5M:r;>MIr B꟏هH|3~{რһ4S%C=lƵP*QM1DG\IH9%9ܼAz' 09]*Dw6ȉ,8tFغf)DՎKn/r҃Mvk4.P̉*QpqʜBNefhCY"'`Ϙ]L@ Я@/~aOS1\E:F$hrʮ-U[j-W!uTaeF < PjJ(i='_[EQla E` (xuyId>P(7׼B!X > Pn%pY*n @k=k..Cǻ=̇?oϵ:2\UjinaɎcW˳TL~.9&8*3StVel+\oE'EܣxLJG5gj1aS=z+Ʌ}kM$}ޝglTJn0hDSgKk|m"~b"fW۰4@ܜt)֘xUPgsP N5XEsu(c.q^wPWT.{Hhޢ~!C41*tSMԀ0Ł4Z1Mn)?Bּ j3Ͱw%MnJ8:Q2 O'bmʲ%˖[^$^"UdjvEhmQė_/E6kKeƬӃ)kQtq k([ߴh.CwA=W81N=S۽CaH֤Բ(yx+/<1H?v%}l8l'vvt F XG1ܦ)% ^PKe`Pe.{HI.}LEG4r+F͈w0kJTQm +Jo1LG43B.kb֚uӐkɰȃp'huH)'U Ե]^#ea⹄*DnM(%B6&Cd3!Z#! GF$!Q/*`?3&ZS_rNΛ6_cW̰jcLE҅#^N#^><N[08s5|"L8ZI"{IN=Ssӵ0D=*?J9@oI}& u)wf3͉jMLZ}8SZeT {>ڶ=^&~Ήv\+`ALwx@q`X?v ZX.:CF \꾇쳯P(!`5vgJr У Rjv/΀Fs7;P| 2VSoթ~ZAo?}*L>/Py)wavIj8A4a]ZPq%X h5`37fR#YOlb̲t-C(+9ԼŞ>oN^qg8ϣcFL^-"x\-3'>:GQ wr;`qͶ}lOtJxx>tJ߁K Ǵgg,'rǡ`»mAбC&&?ϓ !/v82S  SgN52mpCGf6ي=D{frtA<쿡m1WsCioMU$.{A ȉg.ak)>Eva{UqH\ 'N\X9^3cqqAyg uwvRڜݺs5ġ-DǶEAlX9LJY8 z2K&&7 {='#Oz&fTݡcx:e S]{ۣ$tI͖gy2Yx(~h 2ܸٙGjٽ$'b b`wT~\OPWbMփ[a)0` "dRޖñt/ĉ@ 4L&qV $\ټ8;T~UO•[W=$9L#˨~ּ(tQwAQe!Q.Z{owcƬyTm3e.%t~送o[2wcpO xt$wx|Vo v.tmﭖnGn?vtohvum|,]\ַQӍٖwC/;{toлqo{\|~ta^v6K 89>vB}|rCN}4 2Dřڳ%;?SvCzq{?lnc\kHL*Ψ'@H+oE g{c( =DkWχ>pNrڗJ|tOM3!k/psO3j42eqoneI@o(@ kXUiKg d{S^Vƺ2ЇgJ힗rZ{so\4b sŽMqͽR=Wj/=֒|76axJVjIWjP;[5gR{攃gRM!|v/T^Q7: @^d Vf;O:c ; Rf(٨ԻؗҬ>3$!Fyqi!]K,^(.)~do+l-rTSE[^KΐD_s.:і28Lӯ?~_ M(C?lI 481@b~VtLCBnVH}7PXBk(k?䂾fMeԅRtRG0s!I Q<^"w/E>gϿ~h  3"uo~*Mc`1D.8d\V :K}R $ՒI+Dmu5cNiK.) $p2GG[s2[Lh25|2e4pa4o8XpvВljYg-xC2gmL:koqF @1qPSgC!s! 7+BVIJM 3[2{\/x3gS<.ou+\g̹w>¶Pm4CMwMsaV阢-a)}5lhLc&V18 h@6$CHwJ8rEe=e vAek78TfvP\y p̅ bOb2e%P{ڧ\m綏^,V.CΘ#wX&[~gY/j5"܈>e[lT'3l;d<25 6kߜ8M<>;_ FG3qpK]\L qdc;Ay ^KkDe YuG8*?s6Ƚ7)fK콶:::g!Go,"/qэ? Ёݘ9' +}w/ɉg@DYE*sWl 6ݍm=Oteb0'֦{IN<N&E.MS7S#}Ge@#O\Zՙ1>] +qI5e9)38ҜWOf 5E%>vB$ksms5,.uM$9온d ?MwcF3v6EԸcp%M`m-&蛕dANFq7?~9_jbWVlL˗f.\y۩"i8_wywP &~؉QZӥO]\ w!Cn"r-xhxGs>Hr1D!\ta+F m9"zWDM*ؿ(xbp+7 ) ?E rp<И/Ma^˿eu=E=}=[~YVgQzמ'Nt^'\~;z\p +[ G{wbxO5 fjq` >=jb(<@p4Ƨʏ>xseL\8g[#o6FhVjtta=!tf7ˁ#Y6R!J`oC!x]^!||K73&Vx`ϷbrxU!loD!zXC0EEF\`^[oT[!#1+".tN)X^ma=? BDU ݡfRc݇{9f&Vfv,P1 BOh_l(As=Z%Vz"qŴJcJת;INUOWAW;*e-6FϔWtuJC)FR[58} I&PND#SpqIWtH@Wf™I›^ũż[C_devT=9o>j4|UR=ƺFtjB55ku0jEme0xVf2'(wʚz&_?Ïnl"!AbSѻD\5W:8jt"++J5:d\4&Ou1BTkS 祘ڔMɏX a婮>"F5VpÏ]|T:XCVpԫ15fhd뿂r1]/ZQқif4v3Bv tf  &TJ0v9U~Ĕ@{n/IEq 6{5Dv]\UFTSC !vBY !֨^뀓PB0iږTZ"qM$ #yzIj-)xh _+?^S.z6撴?3fT)_( >2!iЮڭ92fF.0P"@*ӂ ̡QEE)ߗ.7e0Bsϓ.FbZZVŜF0dA$@$4T9whlvϘVPI|a" ;X]NrFq]ÍU7< 3GvƲoGmv9efQUic4hW%ߌ+lގt(",DOJJH~x?>E !b)I$&M e>*M(q2{p0CdE[4d Fy;N Ã7 JZP66T8w}F/y ǍMLj(}&3EU\D-R.7+gP"y(q*E=B-*T#TT@N'P#)ߠGԣVGPS.lT##3朰 CPIԣS>j !<ЮCvE9T K+@/hW%]Z퐍F_[%j)$ hӂ:h92ūWt|? m>&A6p2cA V`q[仙8qLQXtdX|vz2rr!݉=:HGCdO9 Ll?HD$~yi/U1zݫnӱ&6{d*> dE@v@. JmrC2Ao@sWe58cŪD JmqG 0آ 6ˏ^| B Q_tU,zR)5#P. -MuJ /]^GU[ZfH]r}}13E 52%qpȈI8W6d[Jyb{gO"?O(yf1(\^zo]<r0ã{to-R*@[\P'|+λ B@= :cP)u_:{-. ICeu9^GFC`bJ5,x+RAQ3'0In$G7:|RLmPxBJ*P/2hP*rWDѨm$@a'GrN2xB#J}YQnx~0Ob-n~{u"VD>Eȫus"K 9jn.|ӺO yKʵ4oLA M 瑳1ҙ$"&ů<`.ϢOy~򟏶Ũ-Fňn[jF"-lIʀT!¯ ]1P_%ҥ1uF].ZDryP䉪u$} u`g19s=B t4gJDN卒=ޭoc㾕ب5qmq\4L2׳9"/h]Ս.(pB7A!$N!$0u\#9#m?RiWE4n4ދ>Gr41i} Y\ 6" b[XHn)s ihYӭ >Z+XIsk1f0bkri'H 1U %cdp_x9%3[/.IFT{\5qV@'re a_b\u1opd$3" Tp ]HH%P>u^V>\iUo^['o]N眩CeOP<&C_.l}/e62`%BY$67Q.+*ņ-KcC)Xc ]_^vBTev\4g?3C̲A(D@w~-lmyMwHG'vGE׍9 Ϧ;L@ [d>4"aܱa+ग़u2<2xY׈Ju}ꊒGzZ(3wK:#ggx1qԧf7"KeԹ M&S%8BRgXqeX,I@,\LyBULKcu*xbhdx7?13E 5"R>XŌsZpRdi .cL2LX(`Ih%aVR4"[Ajn\ML#2ӑL^%cV+u4j 5>ݦ.dS|X!b%KB6("ʄD-5$K$Q<˄TRՌ8EqsjߣEk"ӱE"X]gDs;Ʀn: @=+7ݫ/j8H*7/箓Ç~he9?Cq#?:@?7M<_YNYNfqJ1ݾZ\-_1oCnXE4;@]w7]ԋr3qxNb~c9O K$x!-T7﹠c=o5!5H&Cƚɑ*P?)X>[#kr|3v&7d2LnjE7e@ 5[0="$;LamEgjy Vc"^X.#Q9pLg?7碸,!D\Qkցa5aiy .qG Bݽ7_[ 8U0"{p{S<}kbѧC6j2E9d/ǣlS?dz=;5ZRBS㉷ͦTN5-kWO&YtzG.̿[>e_YtV" c7w- j1fDY7U`,8ZC3RRҌp9N󓳈j|N~O<"g-1b(wGٜpP}\̳X$Y E6 xKFFL=0LyL QidWs΃F!(yzk!%eiR\Ty( Ǟq+Qbu^ihKce.QE;-{n7# U3@H4p FrLen0Nɰٯ!BJX/4#2$Ww:BD>\S!OS9ywvi“pŋPhpa+ՆM/{(qƄiʃ{-Bte5ya4o ?Ïyt瑱7v1?^52/+5c/\ף_ P% wgM"Umߟ'_qK<~L~-\D|ܧx܍~{o#JHSw?;GzF+X4̋u<;xOT(Lkzmy%9<>D4fQBuikh3([k]b$RFK o<=Qz͋-yC)y&}?wHvXE!|ί1,-RWǣx1mn`+ĄoktI)nJs]ʺKIW-+o~}/ʼnW5w"Lݬyipw힥ozF;; ?6ŝDiA*m5_?+R2δ\84k2)BzTiTq{$I<|owf<'M#ے(k-V-־wښJ#*G)e@$IIq00p"I3:g&vMҦjT,Y-8Rpĥ╷᝕}XwmmKz98~b$/' o7")jEL)ҚQdr/_UWWu%7 2e?yu11>Y#uR(Quydrj3O45*]FtxlXgt:8r;9$Bp"҅r.+g}D\{ri%bO!_o*+3S20a~~w!6+RXYk2Ϫfb* wi T}\9lU=:X\ &jn2~GG@QgC?4IOI JY{F2TdD{|AjCREk+"S[LiVF^STB R)X3))=[ Eq/I M`;̕+rsP5Oy OC`yuhZQ9ˠ.-)A ˤEv?YiWVR QTTL UJD# >6_c @ap_<%;E6~X9a{3!e$W&jA u@YyɆ׌:hNt*Rʭ2%s%KeJ?Zܦ.m^t>RRaէCftѐ"2s: ڂ;<Y$eE:Uj\/Lx~gO $K|CJ4'F >" O97 p)2W"w%äXvGG4|Q$ !7/k>=au.Jf\C{uW&kDԎ#XZ`,5U}Blþ0pa8LWtck)](I}ux\7&TJ^ΪYM""jޱcH"0+]KS'+S3Za4VGn$[:O(*BceM+ _$/o\BOE%ֲ/Dfb'/⊣D&ޅ0˼X%/lB-~ @0a `q=]d׾ \&7W㜖a_)FĿݒx+ubvX>q8,{*/PT5+{)94RUʌp*%FUFqTAfPDGc24DF7m|L VqqYdjwBRE<! Hjt ?^5b_Ԑx_œ'`_8BJ -})åWw?NBф"qGY\*lR\Qփ9 (`ZziU5ŜU\v-3ORh^:tRlNLFIXQk2W #S)qh[A`G١rniSe4Su&y T;u3id"8OqJK;k3N80X*AXXT}#W)GpڤGU=Tr" 6P)!BB`g3b⍚\ͷGYpu79X6!NN 3@(j sԉ#Dz+!q Oz dw5=EV?yװ{.̿K T"$ r^Kl I8Gc[pDRwX k4?_9 qm稘M𝳖's}p.ɲ:Y\׈`u-"f*K<Ĵ2_@\-f~1:vU`F23uňC{_]],B3 2[9qAG탒ES!H Ń=fWGYy'buS%`"dNfdNfdLZ:ju0)k BZM8jj:180&.YVQQɲ[2AR } ^SE rl>#QIPx<^FDbDC=}aW)u(eIZxq.5IرO \TXό6Eh{DDDfPkZIT[$&Z9CZxp6xbŽ۳ Y׹Of4@˓'Xwܙ B@.1ALʺf 8x&a8H˝P͘.0HBEgX<ܳn=P] #wҳ" @.jSfa B{,/2UЄc/0љyn`ّ½ .v}'ȚC)L7=d%Yp tGO$) (MM N&v~$Rsz}>BD9s%Us@zoKrZzķ_IAB/q?GtVŬyhԩbIxz Ok~ 'L~V>m<ʹRo694&̈́.*W` o+ Vy0SO^eQrϜ^7xJ=\ z٢lM_w/wU?Ex aJ% 2{g9o7[}g6~?Ó4RKo;>"K~tK?1>w$Ђ%xLj9ZKj~!g2>^ڜCֿ!⠦B]Gq!(ǽ8Dq4h G92MmdTؚ0=;u]טEX֔pe LloҎJ;oͩ0ɾC7E(#CR80aH ü JCJQykj'ZCݼ7RqLw.zB\=q6Ǘ1]tUxviZ,'ygؘLcƧ`Q&c9m%35Lԫ cΩgwH5NY_ P,` 6us-EaG9"F||0(;;X_CYصXC;{~ z5ԌxܯF xhYF mZT ηx$ ^Q$ +9˘+X !xq+4:a1 řbP|gAJ#aէ?wOb[f~D%^ 6{!. q^؈4;P`|}9g iesQѪBJ+ 1I U ߑA[_&<|jSF,@]PBiǗ=P LIvmR-d Ď^`FN1\ B/k#hERHK&,;a{,6k!VAvЯ!3T DQ5N}Lqg*WY>u,\ٛGǪcogc:?O#3"U^1`UEDDTKWaϘºڙGl9/ւ^gbF2ϋ~#>຿M؉ )G͌72nZReug_Z? |<R!ٛrY~["ax=eO#AI1EmZPQ1%TUK+Q0U4H7tLfSϿJP?nzn9]^®#tQ4B{&.S,x(ebf աEu%e1Ekr!Ң\~Jt3͜4=@d =ZYЌ_. v,% 4WOTΝA3o|Q*Of39#j=F q6U0"1zp^z3F>x nqsn7pJ3nҊqY0- 7^ ltvXM5uXHjkAw23ʫ Z[8*PT:=0rRԞc!''(՟l [iN/[u0cX@g$vT8ii]UٜP-@QzS31 ހċ3[A,{sdu+0m? nCǢqiP*שĨ52{o(YAw<)(lRUzUfU'ncŐjKTYMik7zA7oCZnpA!TÀ0}+p{hn@@2㴸cP- xuvQQگDkR%4)%k ]UȭܴM,0#}D | nw]+;fZE3X,\ce$Q;-Qb&ŃlT#&Fz9`Ami]E]آ[=T0CjTXS2AmuxkܗZjt4]sNwD n*'&!O;=n$/>owXSxfA!x5fƬ"L9C4ug|dk$LL%H'>01 7&Z~` }Kv(G q=!g6 m7Pj.׳ZN2v |4+)eA>A6nÔO3fpў\S1h*q٣[$gڥW]y:w*W:tQ2wp2ur3FkUGh "\RƍIT'`yQgWs$鋽fM"Ez{OIfKklޕL Dōv(s\qO]rnA*ZM?d!#cuTT3C)">#0Og0u٪ :56 g3: XrwDhvO߫;~2=E̍#pz㓢s.zRxz~19ɾ/o{~{w[x:[_#\)|էALϑ:bp1/{4kן7zu=?|;=SP9nAcl__NS3t8T:OY= W><t6t.AAޏkXҟk#e˕Ny,b`v9x~;eb3K/47~`uOǓ4['̳H4 _rC' ,w?xO~;ū#3cb0r1Gu9z2(xFa P~ǣOW~w:{ǓpiԳ5Ü?+{ۯ '@?A} Ӭ'_LtPddz1};~=D)}~ r5'gl)*ހX/f8%ތ_f/G.fq~8_}o2Wy^ĸs,FV\ 1M/y_v.~S,!82 )8a$Ja1J1~t%b2ɷOa 8̥a Li)3? 6063K?d 70L[! ǠCf.{{od1\*3>UGpm,R͕ U]%@gjUv1|,,:[SVS_;\.v3 Oqt>:Y3%c ͦgLŨN$t?͖of<;};kRz<ͥR\&pjcG4*:'8q"(ݜy#JB޼@+PRFfc{RZ=] ֊jW@KNu)^*N@xaM:@3A!;N3ip)\:{i =GL1q,b^j&~zwg7GeB1nSRiXإsEQ01oQU! M@B?$>i4 ð|Ā`*/$xEl"R!3Sٖ< Ψ>4?wfK {a5Eڱ i~(^0D>,ەv> d uU#lYwFVB"SHpUD(ŗF1$EJvpD2ƟD28󔆡uFT_ )J %(̌q"S"i)ܘD +- A"D+i┩4}kd/[q)f6U0ާpy]1Ew h8W;;hzgS=hf!pܭ ~P{r@h v>S\RN(tA(Et(Bw(]Jr[7E'39Lb1`aDB!TdlAI%+]TfC&K `4(qч/o[ҫN`)^qSLA)?)ZK c7T6z ޏ IQk\N9.`$1C qsvԊ mQ404J}Ir$\v$ 0QMp_kJ+rA9|^\0a2"],o߾mw?gM(<_Eg-mVYE094@NXeO.t=²Qa@ͅ)Ќ 9_$ir'yÿsll:CK}[Y%h6fSkJB,}['OD//'[ jc71hH}Imsjsiw[DuK8q+ kEzFu+  61lIxpw5OW ?b04p0Z8O ПDy^2r,<>h'R9)> cʹb1oXδJ.gZBՅPBd B*9ӆ+EK,2وa" K$}"% c͝N2]֡ LLGDH|nnLHxR0[%,6yh|h*# V@r2eJr> IB?#hꤠÎqIb!1Q8dTJ3{9L F}lq@|J>]o_k,gXw_3{<ɧ+|w_2e6SG3n5AxWe{ ^ʼOj`aPmDPsR gJH!(2%iʱ%R+G+AinjU V=h/˩d˭gm` b$yv EB% NZ\6pήK/*syRq&A2eY+8M}- ؒb5RWܝ)JKcxƎ "{;aN'NenbT'"cF#:dc$OximS6Da6m7sSi&bj?Pllnءz] \lRaDZ,F$T+M M@Q뷭d*Zj IkTUh,$Q>J!Y2 Ԥ$Iaj˨QcESb8!k@Lp-)LhsM\%Ry,JY䕲*-R#T(aSXS~j0)fF9 IVN&YjȪnmoVR26262626**֕Jzm0{JDZQLR B!DL& !&[:-rMdjcĜlvPr$NRĘF^@jkDR3!s ԟ8DR sp86A @2Ic$BgD7t|qq9̿ B*A yݦw}@ޣͺMnӻޞ7CEflˮB"F)Ġ1!NQTD܇]*BRzoj6y{!e f(u "tNt.钝.YI4ޢKJiGc1(E(h!1}%ad6yd|Ul&X`'T #CRKe$.\,D:đRk*ЭI'5o% 7Ҙ3BP壢`W6G{T+,Kx[;Wݽ)0JdI'R߯1C!ļ(JKyXgF?ЍY>I^=Q?x޾y:%*h-qpa|" !?œɏ?|rtu?_,d~O ş !şw|2vH풦=8s+V$1"J=OØ]^I# X=CĘlũ+lG!Sn{r'kc=]`du]ڲG5}=,^c@ aAɌQ5#3/5p!Ň?hk4Ydw v%zEy(9}F,z?B4u>nW/CU޲4_}'.1 (2sUP-~A-{M7#f'دk>O<,ʩsk>5Ycƹ͂3=jG08M5Sa4DssOafIm3O:nSSNP|wieLjNqKX vD ZG+Jxa$b$bdC TNL(nH?9q j$PV|t~&נ=ꦵ&eM֋TT X˞P,OX˞c+y/M,cobbj1/+,lbBxmycC.("۴X474(WoxZo?Wy͚5źOkA0TOE'){}*[JXJƊV$`Bm%wkE* 2H ή'F3ӄu^PR}h 1Fڲ>Ay* b"S"7౥,;E36e| 6I/RGn[x$OC=4Qݟ_8I4nxIh$G}<вAJ~`t߮} 6߀T]v1ÏDZ^@㚽?;e4jȋ啄 !x.P'n`~3?5 ~)Ho|w!+YsJL/JlM} ;f0|{ <)W'{#{ ~b,\YFޭGiW? (N[fVO{I]>jg vQD+ʆUǩs*H~rs[g{ٛrvsoJVF=Q<.2>߆yjx:~J 1&+V3{c:TE:UYy{L^Oד#CRl0hwluͪ4ZAGEogukaN#1lTB#v>:wOc;WT"m"|8TlĶ>{Χ3`Z^: a f$Ɯpz&FKVP?ARC_>wuIQ֋J&± Faܒ1$,IF%(LjD KKk6>,~G]!LZ̖>8Cޝ!=arvчzp3s>]eGy!y|G!]Rp[O_I]·+b6OM9a1WHvޡ_NՓO^֝t{\cɭ4er+iV4%9ӆ/IE'w ?%)P6h;Tύ {_Jo6}tzDOw;F;~|A0*٩u`:Ȍ0¹yAm*u><^HZM1ԔLlk%34_oS&:+geKj o (n:2ܵPXQCGf,bh,̝PҊPjrT@ J t)" p. 66#G`o=*@j5pqg'lᲃgQeM9#I@Yv+8y u79 0lALB&K :ekZޫ8A { H߅ hHeD*\ shI(il$Ld;IA*H]hXQSޕK˨P6OAPIq֮+it^*__/bTp|w}"103*8W08Ə7SՖqe(ƝzyqѩeXE- ]K|o*x}QylIϯnݕAA"[^Pe g`3F[` fn3X n )g9fpYoN\qv= Ŗ\8mZڇN:/Qntq<–kAD2[]tdg]Vre̥ДLCfçFI2rE8Ƨ9mL7՜>Gx<4QNG9s1A*+Tmt;O^b\kt'Jq 3x6T Sg:Г c0J 45!?^z$*xc,S"c e:kzjEC*J{\er^'P.71՝dg%lJवLraa}s T{% EJ)äc3N" @#E{pt6(vR.F<|&HDT*fd)L ʧ F*j+#K!kg<~K ok ϴZ)yhBGr#FH@G# \]MhwPf1jdA$85o3Bq[) 8wdh?JFhƭWwqJ- "^J R#XR6%Hέd >BZG&R\bs ARIΙ`FK7V ؊x_LAǾԬH`% v#T&-iY!픵6+}tPy-v9gȨ!*l .a4"7TE4,jJm 8|(u4X ~d4U>J-Io*ڼY#6nj c 2?K?k+f6m2PQ2.s[!#<ÿ0Iɉ *t̼B"Eƍ<(ޢ" P ƕc)*.IXR񀱒ʠRjB1kz gLa)~݅#J`F2~w;rAaQK%Ij'|kK4\̔Р)pZɩL=K!(Z.eg*Լ^UFQk,K_eFO4.h=T w!z $'C>if) ]b{{gBvhӄ _Y-{hpʛdrT3BdS3-4e.{0x''TKۖ=Faԕr:VGx!B-E3VV|{Էz S|&^sB?ltŧ 3P|PQ-Hk>BF pc[N8MЀֲH]Ĉ7H 'g"€ KǂeVX րR4υYm@9hz!>C?nǶ"Gw7#d{_=H 7/i3!0c4"`A9bh<~VZ=-2  :=3p>S4bObb% g5D>rA܊ MJ; ״LMDt-]CɁK[O~no%>@o.sWi%\\sO/%̮ ~0CBDžk\&tMl8jJglYahT7u:FeocJ%4L@8 g hn;ӂRVn0tΠgs#(d9@ةt;:# 6r鵁d;9 G!ٲ0%D-agJh@SΎoN`jqęKBbVKnՅӨ^\;%i&h9zQwa +o_N~N&dW=ZzW7:ol4}|ܘq4Xp#SƘ<צUa#(oL3`?ˮ 3TB/Q7 8WfDby{%p"Oyhs(J'<I蠅 EGW」4{v5}\:_T }R5Sl9?]B"w5=6L?B]qTynSJlڑ@4~^߆Ղ$QwWb5{୚Q¸ pY"E6>B| l4| i#! :q L୧4<^ܴJ"v;NFF%8` _& n\z,jŇapPxE>q9\=p2GD,~mdy%93`USMIEIQmͮz G:֔*LVjM:J\~d^袻mM\o(S˦MD#r%5V# u/H=]5Xt h`W!S;aaǪ l?3\rY s8吼..j}Le]q[pӇ4q{Z& PKa}=@ShQSh[CO~)(Dq``SMVvw⽇Ik-R^!,F#r!JQXK mEq[yK-n0BV}Vn>|mEw<9(:Oq$"땏Ve>ZYKi  C:kÛgVI7o=d*P+;.ny#˫Q`ySYx2T*1{kJ/٬YNҪh-'i=>eUߒ > Dw:yaJclM~ǧ2Ưwij9uy:\kVtS*t!.姞=YJwY- Mb7[Ax1,8-tcpLwPB-dDr8ftڲ|݇cJ1u1ch&aDtK%xs˧q%gfbvywQ*`@hk^` "C2Ad& *K^M5Nv2GÕ7*CAN¢P)z?,:'b]HnjIX s?P:&aP4T:/HR鞊` k`($!X#kq(^q#!$lB Z%B!VIH#߈殺4^b#4Y/Bu2:}s'Y N0,idlI3 J݁+W9"hFF[Uthl#T"x(Ûr޷F 'Z#xi?6i~' 9bOʘ2M@o ,93TuzJ+P]ynB0u1J'|S0.b8mu-p譇;aJ`ny7-XHv*2Oq7`5N5vl]PiT'%TѝMD^G0x_~d1ĚH1kU,LoW;JlQi{XaD23Hb!G_? 3oǞ,)Dј(^01˅$.OةD0juWHF>i֊Ayz(:Q DfJQQE 4 ^ N l,w)2ͱ,dS_{q?5Y,JӠ6 DAgC-3RlHF&$}ºa_tpС- jOm8.Z̉QSLT:3G%f (a֚ !265:v 䆁 /0"vPI5VGBb BY7cmSdkb¶Eu b5JۍY•:?!  NƃeA~&2xWg{L!WM?DOkn0< Q/gb??<$0Y꾼Ezy$lI<>};SE; [?·эݸэݸ*F7 %6*#SLy5Z+ >Vsi#LH }#5r|.`DΕ\A.K?$[RqR_oº/_RQɹ m]&!*ѰMŀTpĚ`0`\q zF˥V!Z7bx ǩRc0 rY19t6y4轳'Z2 ,//4t#whFfX7ץB6;)Z>=+,2=sGu}-y|a 3!g|30`:h&|BlVjߙ,9<'X67~<~՟ WLKp?[x*ƍ19BպъF%WڮDfP_Di2ɿ|p+׬po7.>^,_}JJ/N!~% Պ|D| "&XW*C"FH'Q%nmk!ҩq%ƦB 1ޓXB[uQ1{Ӽoq??y͏~q9ER. /OCΜ<Idw릓ɼs>3gyDe_̥P%;~s;Z6J/]1eūJJӥ,Qף5$~s uWc#Fݙl蕇&fH"(ne Gw+h­*/+ܵ.U #FLIZbYɩXUݖw|H "s+|?RbR)\udhzRdVѢG'l m ָi/:kXx*Z`'o p@Nʬi=(v 6R2)rrK U#՜I  4tB! YvKk V# r%lm{-X[[` 7%AyX9ibL39[Yq;(D|=XE9VLٻ<: ª:uQ#^Ex5UԈWEXLI@a,sR2Wx^dkbXNx!\UNH}QR@oASȯOm\[@L X8&s3yY&3!=ȞDi/7^2Fh7T*_xR.X k^b3 ཊ)T˘6(0q$×'U#kO4 uU]k8cf:E5].AϙR1ʳ%8}%"U%%8X%6Gk B՜,%xK*f zHJ$) &ArE" K6@`^{bmِ:A()vj2Q|bၪ.q0A,D21G`^*^Q8$ٺڞdx8hgzp)4渱q3s]v>A)F25[=YVy.3P܃6_g+3}p2f rexPQC{ƐG6vܕW6`FK~X蘳Vxy= 2FZ0:ޒvƁHD-lAW'q{-+v~>I3h1eHm {cƴ)03`YP1`>;@a JaFcATn4 spqRn4C8d(ؿ+f]p2#aK>?0U%EpNaf4`]HHǭȠ3 S7y^'4Mo7f bWwJ_>uwϕrZd) HluW0^򩒗~Cf6y)J5 ྥQuGIox=0|cjJ" T!bW@a'ԡQki( qXҀ(&ZZ:Iv\:`Tps#Y F&aDV0'%IBky#I.*]Kٻ'qcۊ}}ٲ+ٍ|Ȧ\ ڪF;k 5uD8\Et7w_gsvw߬tåC?ra̧ל>}6sx(TPKump 4 -;^m!\ tLD O4L*"TI4V+> 8R^P=f+3J [cc?@W+pVbMb@m OɄ~¹S~ Tkh<&wd66n݃}gC?+wKntM.M? "gX :Ey[Z7U>>1/XOCzc+LD4Ξ<*EZ1Hh ,^塲!Ɯ]xkw0P*JJE@vRwiˀ=N8~-8l-i_0VÒO ,m "PS|"<_#v 6%+<˚إԊrt3O.^; *VԳۏfv Sx w=52\1AsBG$v!sEyhq Qίr&l%.Fհ%sAC 2J3҄ZvH!0.GLU&i@+\(Ɂ>IT.ø{Ɔ S $dtyw?y:io:``M0uQ,g977^.k\olo# h[8fbp}e< Xm Nw۩)eB7boWwZny=Zm90n'}ǟ3Zh&xhl}CZ| R Hݤ*7}loA9ܡTks}1?_bC0p5W; XCɧd qh}HKޔDTH#a@HZ̸ RcB@+,EbLLaJ]L 3\2$1栟Q?0rqNw/]yҕ;)@h4/Ώ;VvŞ1vdƢEzPh}CcYx 0yȓU8hȴeH%Eiv#6fDRws̬ў+c$!I3b"a3iLUo#k索/S, 4q7Yn~gx|hXңjpO1QFR4Ol1HF(NN`dX"#ad8vc`[)AFTGGv>G-ܖ3T/wGAQ&)3z,gsp5w#L]v/LBD).8QuB&NZlUB.]W^+IJÊt9{ ߙuKZgZ񘉹516HXÊ'%NHq 䈔^Ċ" F{P,HJ>sG -DRDLّ]; p{JUx6wf˻w sIw˯Ԧu~,O<-|ȥP6xzZ0 ~"u\(&j5/w?O.%ɀM>KB/ cxӗ( {n,= +iCʕT |,Z(!7Ov+A)Fv93:Дƴ[nUpW΢xJֻfMVhamns(S"nVzy[h)"!=[ᙖ mv\XhJ1kLh^V|,Z(j}w_e]P mv`hJycڭ@j*8+gRpF6>LgsI~RBuT!?6b:μ`+Or^s$Cg|7Ňύ@2Ai5*Hy|4Sm8@< 4|Xk%U%#'%Sid7BA-3$ūJΧqʊgEi;[7^eZka5BDL_ӻHCoS76ӳ>VI <~;I>S@&H͡v4L//6,.fRѴnȳfe ~C{ x]l44Ѕ~Pm@9 BA=gVj*qIVUSkti/;7{Ʃ96[MjM'`Ebsu9T80jRMꔙTO!ꂅq^M*fiuǻs,(F Pz_(BCm7 3!!XCn{ugiͲm ۿDLWuWx.~x5C;rJgj~ y\8aw) Q4[氭vHlڗq$q[I{((qb X}d#?ᏽLpwM3H?HV&=PfB| ?PdB?6L{<p{ڡ)U-lt)z ½աJ(sRZBݽR͓6y ޗd6[IR 4nWͦ{7NjĠ)B?ƽU Uh1N$-6P+XN\ǯ-R ZT+Z3g u|oX[ Y"`c{Cww_ga5xwd9R4O C A60 _K9Ye.($=vVU0pHIc`Xpm4V+> 8R^Pνp Mnj(tA~v&( J|2N`ev b1@L'Px: ,s" یܧ0%qMT!{T8ͨ%{porD@װkot|׀{l1`ն6̞W6-bq e`y.c$D1F؁8 8G^J$T(W"FRF*5&ӼUZD"}JkJ DU&3Wgz-IGK2QL!Hv2qQ T^ ^H ,7R|8] 7}Ft2]vdR|| 4# ߾ygM|Cv\pv.8,(9|NN4-qّGzZ.:ö;iQh4VfOOsvj0mǧR3OsAhjI(JI;}]kSpSp!(SpSp$1z5{'ԁ5 5}6 …o]#BDt\LjǼk `j'ɋ00*ez;(2 Jmd[/R &mONf*XoTC|2H_Ň:Uҹ@|4-g0h%"fֲ-)VR"1]=lgI -TD Y8ͱGiǰ68LFa N{jE݃$Zd :o)!cP%!~Hi4lA ]l,0̵ ;2sDv 7m۳ɥ0*5<%;y*"TN#d (錳y,%eG _T.ϩē/,}씖O Uزj|d d?u8 '?QxP\\+W^ !!ƮG K~]9 "OA;&OxjdR8,BkM"Qb!bl䠍)ӆ#i%S2䢈z V#RECBcyi"H$uD+ktǒq1f 7uIb0I`EX|D) &OB^yTF:{3J q$Q>bf6}r*RW:Aw`~ŗۓdF(S=qp„mcۿ"%a|)nH%Q~+Ul z 0+6Հr$xķnoEntW6mo#)#s uM ήSFҐ( q_ @։; eH"PD]KW s! P$G“HićQrsE.w9X }ڵx. yGWhxuOդD6kmQv05ғܖ>Wdg4FI(ј6Z=#cx.J Hm `C@^2΢q䅾 ytdE:p 8sFmO!Īdz*@iR+c!dE+_W7Z@=wq:':AF_͹8FIA`Yj:}.8HsSdZ΃lc}IxjGru#㎯ @:7_'H xonҪu1u0 5i)RI-t(o_IEW/# UJکu0bvT+|pEs&9SnRT^[zH2(|"G٤y^sgͤ5s^ KCC#A4# 579A0s6PVp,NnۄAst8e)fع5V9?]NQ`Ot1'piBH 'b> _  bؚ݊@ug-&Jdg5S@Aώ?z~:q$~S8;0́] ]8[wu %ڝ|9FE?!?v1@1@#(6zr6N CIs S+~՘H=0\zs V`6)XXs-{w[uRf&[ ZYMuw3qE2&#D ${yqU9ٲm3(VZK|;UGOƩgK&QEo#94}lHAz*}0!qT-JI.n?"6bhdF$/}^F; d2BW˼Qz $4C^ڪ/\+\# ݧVJs+M'W*Z:IZ'x8'`PU}o&]4GI@;eRHI-83ˬy3x٧hxz,Vh 6իwh i&kuα̨۞̈uzpqE9ݖ&&b!C#F#JQQL$] zHSlH%Ր-lޤ*y/BRtu\*d 2<b{dKNIG*F` !I3AjmB:Z?6om6hjqM4z=-jWUVz wnU+cceq +gcUXV{x8;:2P$%+l!F6?Zc;qbe6-ftbEF7+ -%jr{;'CuGT>}u)ֺ_|x%d$ļ?mwC҇_QupDuZ`]GRN%.ZKEOWu,̎(2Yj#ޭCCs$z)\{%<N(&G H)󤞊T$@ Sz`_؊Loo{de5ƌRd!%B 0!=ԁ`+/Om0dZ7akx3eC̡N*쉥v d[NǸuQafQ)-8]T+nã[ )кQz^VV|j.!$DXX$XKH\rlW|O0BLֽ.F '3[UpQ|9HV'Ԃ+NC]_5S=;9BPi û8衬7xӖ* sCaTːx}_,Hoǀ [nu fpsՈ/3I?[,bѓºc:)[licx|c}b(@ ejE-pG-x+~4@b*Oj\5Q2aPzU)6 ܥ:XPs2$~K& G bKikä-Uj-5XpslO~όVCe[-jٶ%oU )QXIM;XXЧo!8`mJ;/CqP7IPZq7J>'9HDߙ@U%CŶn+%N؊hl=q}$KrD-TX]= 2&jP})J|7A#+-@VA 죅{Ws~o Gʞj??X g1v/4̾7G`"3c$  @j@N_"r]_3{JUb޴Ծ'A eYgR:ִYjHP^1]!+l 0Ha?4: Q[ S$#ֶ/-/>R-H?p移e/~ldҜF XOk4T$QW33½^vVd@l{[x3 yY47^Y7A; %$1B0. !!#p]` }zHRSE+b`CglFO9 xlRSRB !2t_ѹPzpw؜*\e:dC{_|:M 0NGc˫*A#Ąc7u)ZJODA\0;(1:*-\˜0FxÄ<*-[JPr*Yb[&*AVC-U-S7*Aͬ~So[E=߾[ɷJ]N+b,T5>V)JBԧL:A$9=$ÛQCα̮G@;.ӓ s3 i<$mR )NIG{s6^{9؊<ÈʨQJإ|ڐ#*:&YM2c5$FD QN&hh%A5Ol͜ms~IWT3,QвA7ssh[I@@O&9+aM:1( V֮3Fo)h"<:DatX%⤯ [@ѱ>7СmR{:p`h|s l(LblWJpsAo3L P`dmcN1$XK×6 wiƔ@WIbH4Bi/ˆG^w23 r L.wc,d_AoiܚO* q57qnMh:t |.[ %Һޙ'Q6pblD>y#3[[Yc{tZ Љ5 Iiת"N3"k%:XSkArLS=%g2VDid4/VF 1uCJ +?G}Aea4]{˻e%P=}9TwÈi|ȃJ8. BF} Y]/"! %bd jA k r! Q3"DCA<u vϥAX`|.q]nA9\PJւPj@atl2vmӊhБ:a(0n2hF0o҅|hRdoYӃ`5ПzCR IX޿[${ǵZg}4T@NWCst ?g /cbcWCu:ٿ.L$lUM\ߧl T`@9^S͑@nn}2#nDdu:]?8T/V2%2+re:_Wßt\^\@0BKF IzPtA% w?|iB=桞'/Z-;ۘ gY~0(A!lq5Ȧlb%0Oa(X5~mv\oK&C*1y'ҘLwbL~XNo}sC{Q`: 6n!ʩRy-91S>osKɁ:"󢨛7Ԓjx.. ;(ToI*B`W\r+13!~/R+/K&0A4Ny(m͝V]CuVSPcY+3 DPn7Ԭ"Ljaуųl( LɷFCfQGW|Xn9pզxVB ˻AsU gm~_ _]MӧJdٕ:Eΐ$Rj8(D86g9s.a<M̓wz:ȕ}b?1p4CA4_c,ZomZaCDv0rB: F6~hqc^wvvam_n*Ë-ݸgY$ ;M' c鏫sԢ"g ;Fh~;P>UX/Y1rA5 TBf% M& CvB[o۲hxŅ2gs.$ PIYrD!Ĩ"11gTf*J^ңOwS[;=7f*^5 j|0WP.2A v*eK|PBbVIX[9(mܱA$^0/m¶XKHOx%$9?km(mx uݍ^ns{KGwyEH S `UH x`sq(Q'LTF -$S4C"s LL t {C͈@ %S j:' %6cp$zjF6sZX+`׶`fdx`y`fDo:-jgE gTz Tn~~(3{l_hW^"vFd6?}YG $mRE7؁SŅTtl53@Pc ܬRhIzHh(V08x7*5.p밆yQRv sFJ] ( kq\eT\&`JdLWur5rEXy`ƐbBhs{]T OD9D8SlI)t ݖf/S'3H;?dsļr网!PHy4;l`$(q؞`+U45"hdZv9!SQC{(A0ñu0ŝ6iQa#9; KlHr91 'B#ɮ>DD0Ps\c긘?ZȨz"&B( N"TVqյ̉*Kn*dSڐlO9 & Ԭ}"MtkeH>RPms)g56d: ڔ?2N>`"t_*E%+!h1 ۉpk:*j$yխOKjN7&~PpMc9_Zc57r"IffRwoĒv[MpB!x5 k yI0~Mn  G0?O(#U_ݼ#U{4|FE8Jn͗yJ@!ȵ]PzNidES:x>'Lsԩ͂!/[^kP?=wǃ*wрӱjQS% dxyUxx:IL.o3*ۺa„r3׎WN'NJfE~_lܨւʝ2Del[-͕i^'@|ntHX1y(- \ ęPvwEېlYIsZd!NA17nL.rlYj?~nwI8LہtcWUb$prЏٽ9Gf3_t -,-}-f˒Ϣeh0mhpK2e7\Gj55uwWNTZMG+j'w끢%ۃD4Fgkϡ5_.W/G3:!ĕh]F"7@Tɢոb\o@Q}$'24G\$L?NK>9=kZ `l ڝοm&6xc{iɇXP,iyW烟'5%o,~o(]͓ar2Oϡfθ\$Nܾ&|9恎cV<{p|,x36mzVQ#i=ٝ鹽 쥷y,qm yC Y!G\H0iQY0{V SpzuU;vwV^u# Fqrq +c TKo&NGB@#eO-ƞsLŲuz ?> ş^<8(tvx\,-/^.<i$"]#9csR3,O9qRNfPӹÍI?eӘ}P_@z/Mcz4zVQy.Enwsƌd)#EŹSaQe~7f{H#8GԊ؜I8U0IJ (A#ixcG-TbAT* U:cYllN!f!!lsdzB D{TxQŐdy"WPJU d fHȨ P$1TjmM42RZH`FakIbd94K(ˆhOҘ9AIFב{xFTLZѤh#NM>/.|pI:~o]R6ؘ.uWH[w{LZI#?]}p%Q}~02@E4.p*MDq}^h\FI蓊]Gr=,nɯXܖE@Ạz}Ƈ %-K5-í(Z6Hv_cCW%~:i.%Q#&n vk3) mEir%ڐ-fZX^ӂ(~QU[z'av~0i۾ϗ~s@p}Col1]aMe^ӝj0sE9*qw=htp;8HvOTp/6n%jg(<3ġ 8]]qG~41ĭN6cCx(`[ȨD@t zvEؙ%҉arIx!;-9PK.h}IJ⭗-nyˎ]ή\%R"#QarC!"B&8SUe JBJL%e Ҍۤ]HT#fz`YQolv]E݊dbHe^=nBIW{P%b4ODInȝ"T!LSp"R3-)Д0 Dy4!kmm^CiC3t&m^OwjB\6ᦢg6$;rq7nPSƝЋf0AE_G83w͎xy?:2K? -Hw.2E5*-dGմ8@tu>&40C7Op b‹`u%>k 'r  waJeeo߿3B4;= 8-095aa0#_ *+S 8~Ok6KC۟12Ay>Uy8@:0)I(>~FwA6GfϑX %(||EЦP.զ&ͩ.oM=SN[) Kzx59m:xMAft(8av7:D&tzSMAcvC$# u7cEBMތ5b $\۫ںlфFF:"?}&.(zO2$q}T J^Ҵ 닚lmf|ͶؾN4 0^`cx,Ƽ+[g%8UFN~~D"wCP$[='^bXEfz (cqs3+*Ka_U,b͢7$gԋ>{DR̐ NSL[K %D7ԋ v8r]ĉaK5q% >Z(LEHspί:#AZ >Ɩ;K}6+;0ۻX+۪ap?}-len 5nZ淗gx\,'pX,yjyvyC8S~gREwe@yrB1 $t,M,iƔu76%pi)hL@<ZXKo2KܐgY()yBXj dTQhL`ucrJWgf6;[4~ǣXff/&~A_' 64 b̧ nlo߹$c;)5ƬS=Jj ;ZZ>|6*vǮpB4+) mB%ҹ kԙ\ۈeHPYdI)!c2bvB&IV:7bj 44!ѵfȖ,yk|63_Ao7~85(-ðׂd ImC|JG5abCE&Is󋭷\srFlM/e4Pi1VX#"j.U"gk+#I0]'R1S`oBPp4)yGseTs*  v=Pn\ Xw/p?G.] nYfQN2Lyt<n?^\{%o /89saO:P2ӸrՒQ}{o%-8ږ!̸94OP7 x!mX"c IRoC"<\n@Tè))3@j@Hee97P"<(* yE$-Hmv/h< D`D٠oZM&-uC7xk{isIYOߣï?O&kKގoF~o(]͓!ܒ [eZV_wϧr0ؤfұ OuέĻmab3 yOwj0sEV9ʩ=ԯ,yIi}q*G]q%Q,pR>"'+# 9ԱN󂞙.{ js?LDo45 ]i;'뭚ߚ>Pd]WApa >=0' G$][o#7+^3Aɼ_ !pwL7 Ց-$' aIK7QU*0[`Jƹ%=V#R2UwN;*ɒJ7NSu.Tq7*2kQ(C%HykRom 25ۗjJ RRU -c;E>݆"ʐڅR(X}hNAL鞚rk|]Mg\k%7o͞gNnFQm4d\S7ת }\wxpq}1 @ R]WX 5l TOn$ +aݨƺ)'JS"kR< OV+IUTfLUޢYP?rv\$'O"\_:h&=??i%s:|⸽恼xb;WdꘙR~dfl JY<*TIΚc-ޭEn R)SZܺMuտN^Y{>'LISݕnEUUgI)˾h,1!.ik>"̟VR=*#=>S3j:S?^"[&Z,:pxI*wu4NZӞ ^83b="!yA$RF8aD9)3YflM_.bZj#UEX[P0ߞTX_'꾄1&m~oB*"uih# a|ȃj'V?]M@%0mRkȃꑊ`֤Y]dK2]6:$Ρ̊jjq#69H\Dԇ Pux*tS$Bx p ndw3,uGea9G)jBsсs6Oξm-8V3HBz@:{H8͔_@F!Is di!lG$ K$Mb(X W3- Vq73ˢF~.Q/3a@`6Z'mS5$gumPA@v:'TKaS"^ !*PT 3"O@ a.=_b)g&~:^U{ ('*Yy>fo~6{Mʃ1&o|%9m$CױQZY'W )'[EJX#]Wݚx}3cݩKy5Ԁ٬x];aĢp$bdm>\3%;|̓EN4lb@e37}f4u=i~w?Of R1F/e/Qon'͇f`fp4 ]sA}W_0 ^(1'vrE]rؽυd6nt!? ZBoARs4 77Ԅ BipA)]|=-@%?/]̙ůfz1ˡ/’t_x0k&>dkǼ:D歗)lHDt/1ic,:!l7e2f)RܑR|U QVSnͫ'R5Qos-C,fk14)ql8v.X>_΁-@=`KLoA_㕚rzcE) 0W+U`d6!=C%чǷc-u<r7㋍?l=ct 7T`3%H;fk"aoҗ'GThڏ ZŇۏBKi|dxuwOe"rWoC|o\(nN1BU \ϒѸzܶuu a:CZ1c%י8jԜSMtL,EIls)kS!ޤ6lK嵖(v*P]eֲo)$oRK2AZZ0lN5cS6 `pÙk"j+Z i3թͼLζe&%j1UHUQ[넑 DѠLMfT!R{dHq1-JqL.{W:KKDpRO(TK\b!ɩe+*Ցͺ @;-+SY@hxhsXG6#:ӌ`~FGO/cqJ}z86bF]2=٪L**YGJfܦsQX0$26e[IƼQ%l^#؂^`S[K#;9ueg#5(V|ƔǬvcUUUBdK4B"Rjfݑ呓!f9fW_ǼaO̘E($P!`Shey$3_fBAOF3[(Q"5 \K%Z5XiD* &wtg0w5h^`!%"aNDrJRTD7Ǘu,Â&e}rgьf.x3A]RZdNxm0I|M )[:i\I\R0$SusWZ,AĚnr"x?±x؈3 6Xz(Q$Y$e~VOjS)"eR3[Gq4jC45e/ka^R-U3:W;3e,0Pf- upFx 1C_??^VsU:2_1d=>o;~eD9yd9~P-!.˳?3v6fJfx[q.eans^6TE)7Wܼ"] /5.G xo9#4@ $ެp9qļaRsL s!wèx&XƭJt6=--JXN<0VFQm4dB۫ Qft{x9г'>OD6$:!9=T[WҊNXzTBWETI򓋳N&tS8q*VL$lC":e/3T:< ; )9.O) ӯET`;Py%'CƱ*\*LɅ*LMg1eښoRkW _'O"b_:ODh&8 i$ Vcf} jgme>f`+Al}٬U8ǺYxa[s[4swTŚԢ-g.x'xJ'wZ^JwU8<jrY8q.Yx$sՆ e$rCi^F|m-5_ɕl~V 8 >aBD$)ezlisJ[j /9(XGf#`HRͤL׫a;Ǻk,?<<=_ ,4L8!LҰZ&RZi}i”WI1KPʩ| Ld7\hnAxY!<30m 88 ؜PBoH\7L _1 Tk0y2}T#Ejh*$(M7 cD(8%PRCYA_,PqC"?{b0O`| Jo)DjB-gʚ8_AewfR݇"草Ǟ hSeyըDRH/++ʑ0i-IdR њB5scw# Z kPxfInPzwwOMUZڕkw" 5c,∤%\l$$k\Dy<55Qi̺^3yк+g-BZXMv)ШpY:M-vIɠ`Y3^K(sc'0M 0н `ocLP"%w>鰡ВB˕,&|<$XNް%!H+# V*KagXRIrҗFyC5Q* W[P*-/"Dwj!WXI2ph9dKEr,D>h,-/% a$0 Sr{!T+`qI 24NB jLihUY6 Tmh`jcڟ!FI;1r3Ox<\: =0DI~Y 7,h$a@ɜvp)&06!l)qMyG͹xsX)=ŀCOvigYtr4S<ٜe3coJM`՝n HR·ƨ5&yN%)߫DŽ'=ZAz#s'܉Q ƽYz]_}l"{+"ǤKdf||](+_J+ӕF4cM\*Leyqg2!:ޘ9ê|W:j _mYޜ1A3aЬNI6szo!X _|`3%=C!-<{3dAVB5i.5Iq<>cZvj['$\4([+> dc]شȜ5zIS*Ph5LG+-2sVDK#;~+9got(F9{4f!zo:ӽ1T4/GR+bQ~oy%bR`/ag,1?%҃r%ˆQa^JMbX](C*x|t'`ski%#zf^\=ݻ{ *)bIgP-r7I9OwU_ͯ]f˯cćͥ\͖1yd 6zJVe w#A/,,Njr4Zzlx7JTsX*~&MV]ξXge'ꯤg%J1TmX֠\ukIBjM >˃v0gD2Io-Tֆ`˜@J,A' g̚bBU}Fv#i5S(:\㱩2D!"g3v0paxi!-t22d._|ϬnXI:efu~J9g2ZyF'ɑ+A5Zn(tBqd,w[LK/iƬCocηfjC5g "/sFw̓ݰEL9r]s E4Kіv|QPw&:yHITׇ-T܆˳}yd^RsM|ZvߧOyC+kFZ&>-WSħ dm,O6HӖOK)ohx9N]F8'XP1T'J{ z:y(FwzNnA!+6Ts 9/Bm+ssǂAlu2DR"tKtď~ͩ(YQ>7S8+4^N޺Q 6OOmAo}Ibݏ|Q+`l?.kxQ \ܓ/'C|u$ M<=AEKX~}quxٚ{?\]qR0'_ |Ƙ {:[YMߧsԊ2]:Iz"@l7)M*p=CB $HQ63%WGN*lUv\oZ^33^c l<>g}lǼ,W->H1&]!\Kvpy׻$m}pA1F+R`2{oU;z繿U(DpK$(ҵ=3ӎ cYɎ"N,bN+8ihh4T"I'o:Z(ZKK''*0ּ~"2hۻ&~fAn`۝ # N6PDלՖĤ) -8/ ٻ߶,Pl=!Sn],r8YR%97ED;]șy׼jĔ$2[,d.Byް=)H vĂ˿r'v*haQa@jcՃ۫?.cx(u o:~ri5F=k?Є[eC!L%2Uń[fy 5>Z8AA&D:(朂aiȹ8m%d!gR8# ɰg9S 9 #*r|8e~!Uk)?6Y\= 9O18`^d/ŏW02sW^]y1wUZ̥3Q"8Ä~ KkRHJ;trG9!渓*w&JD)W`b=> rLOi_ {u'700l]< ?<^SOI}Qc~}??U^=|_)z|38]u\fgX/:ɛp 8#E \a(Y__at}{չ߽?]ӹ/^/Od7C70׽}ۙ(:ͦsNɷ\>vӝ~{u4y`uuD[`ɇ8VxzRoCv6-;<~?y=s}|/NxvGÛN/~pd6r647go`߷u=@D ߞ & So,l'/`?zff:/^uz__8Fɬn{F_wIJ{ۿhL)/gG&~Ƴn7{(4I9/ꊟ;o14_Կ7u&}6<kOލ?&50 )q.^XK&=z݋d'{>MMQ#S+}=XMu^ʮjqȣ{Jfd<s)E\Y;;0{O4NI#Psg0&/]V:ݯjÇ`6}5kPk PPCR$a/`!sA3Q///w0 F8{ XEr=@+Pˆb($aR9pG9K )SJgꆹn+n&]Jj!&4V(2>j$ ɘ s,ι<(2.SEJ}KY,5k奼Xt![.O1a#!.a@L&HD8@d9Qc0RF0Zi 'OEWa*©WXrLU'K+/JK4WŁC",҄Hȸȷ01Ya<—p+QnZEjї_K6R:W%36+P䁑4 ^TjuKD͊'jgufY'jnJ\34h%iVb6[TǞyr%Y4-%i1ӢrX V${.ͻwb-)[P"%) ,r։P09I0 ⡢D $ [qUX=Ct,-<۴ S{! (i,؆+dT=$,WPP\Ւ+%qL!(*LkRZk)a]Kkv-e*K9JPs0m#AaT3SX>Kr0JtZ~U4O ` =TCSgu貋DoXEk6"ѵXu#ڇ[InDWH49}C%7Q@Vken X+õ2\7<6,< @|pTz.dI]r[?4ũe8r⠶C@TA՚}qPf_kuQ5-D7C2$f=a)qYf)x-^o o]5%4f.g|[-lRKp,@lFM@-gBzͿ;IRӸm555G;NjZ ZA ʹƪOr*So^_9oa)A.vZ}؟?D&bd B!c:í HAK4.As?Kr ck b kVJuMJm ֦`]mFVQDj+-bI AbL%Ŕ㣚J*h>5-\!wmy=w*y,V;U<#U;?7 4m n=X>Qo W_iBs7B'+ӇVR써c"n{h*Sa%퀆U;nI0!D!@ټ*@׷`%~7>sNcB|7|[š6to; 24OɃ~?L,!Ab&g7pM{Re&+ҔAe4P^}x x\qWIMla̻d{ ,wxUE&7؃DJFIգ}^=;XCx_F\ơIRdv%ɞ-N_?-W:H^|+TpTOވ#%EjoR<4#v$`F,EME {!'IH{(B!-PJJ % 8(\C""\F#*sN6 >I+/6f.{<\,TtZ?wL@h_zئiakA|MZ%ea\6ˡHV?RMb #P?Q| $h\M1Af#&vm{_ s_✾MaN l0%u{B~%6:KH6bo>ۋ_q [A*I87nd436zy3vɫ6{99/_$k?փSi%ޝaQp\h]F.v!n5쇾nL+O uf ̊Pu@<&KVȐ9Z8 'p BZ'hʉ䲼~3V&0ǀLВQF6_G48"q/9|+,sbX(hdPcׅo@=-qֱ岄m>hǯ.;o⧰rC(9j-~!DQ}S7G(]axOC`Q65ha>ļʀETp9{teZ‰&%Re pJ⌡qVOrN%1F;C?$X VϳKyF4\;uU"0"p(dG$vQROfZbdE(Jteu~G_~6Ok`( )օ3")cQi$PxJlaE \hKT{q襢Gjf~c;q4ўݘeQo}u{w5z G_+]kivcZ<1!p>v5k4i<0Jju)2'l;qsH|h woϋT?yXĸ,W!kꥭ%CCȶ"8+I%(Z \`F8 dj=ܝ1·dTdJ mHZ'*c^?ϫ軄C=\̽̽ SN#2>)0! dq G-B`J h{) 6ҥMM= qv;&bqbk^ c[⅌QRǡ+_vtׁ7' 61iKS1i<^"/ŋ x=.I-[:5{?o6\g8O4/_T֞?5jҿ$:o*~?5c c+Bnܜau*@Mfk/s?sFgĉD/opsQMKIWr+͠ otmH6) ى- ()T+@Q~]")8BiUʺ( ~] F&c( ccj4%VRmf9ij97 $,j5,?Q(_ISL G,:=8/M%P&0iryyF}m-+3Dń2- 2ID :b$xk(ӘLf>cFʃ^Ziq XW?n#WW^΁ (VS?N;B )cuSS:}!Gg-%7htv:7(PzJꬢS̖Y~x2\+RZLe8 H Rӆԯǯ #R]oǒW}&AZ~ y#y~x}M$ o Gf:MQ=ݿ:*0[ƅP>]xHa=AY >k' r=kg-r WNw)G-;OE#%axY"*$Jq}W\n؏ṲSܯggXr3$3}ϐB3{wDg[`S\F̫1ƂEg &~efZh*\]KFGSOIÏwVYBLܞF˗S.]kOk^]d:y4Dž"ҏa! =2p_-Dž/:Yi`IC.dg;Xӻټo_~4M{[ @{ lFӨ >MLi4o?࠿0ofϬQ}m w)Q'Н_hy/){|պ_MPcB;iɸ}AQK{GmŠIOEvӵcb'=+͎NJ(eUhjm,9.3y {w٨eK䉭ϲ~cL>1mUwo{r0DQ9E{vsKӣrrFpJruygqۓԅbZL[v#1N_J5ԘS6S&CڗB k0t+3nϰTk]gLYVjj-TG:j|/6V.W-nfws7<_5 &GD*vѴbqBrfW afDf&1W1]ş׌ l~ /9yKoHjC,ڀW;iE(V/)VQAԁZ:EIG9"I#[KAOEk u剓P 3;B4G=vs(NФ1/%B6Ljl,H1h"ܡL2˿ )D|o BDq0QLo?䟣GhѰ=t5xd/]_F,h|J5h*j? f`PHpR`||fϓ===`vu6!'vk`(nB'px4x| FQLk)ĹDWy*q"1SwpI4=.UVȋ\أmAAd-bNgAѕ ˂e gAсiTqA˲Eۂ4ZPk\;Ap-XƅD`ywjowGWU*`H_~ n6 0(w21Vmzԇ܍Wy"m^2L8L2 0[,n"7LduB5KΨwR@,Rc Q\)3hSƼ@z^2Qx^8D&0]ܺ]^LE7f~6{{x~vRFԆ%N%\``{]|o8ވPI. b\ c!En٪w&KK0VfM9h5}+Yf^ dKn}ͪKl҂%҂ OJ _dUsfV'n>|(>k,h=32AxE3>LI}JͅHdE5μ|~4;3dL9Ŭg*`͂sJg.F&CN..|mߦxczx*LҢhB3Zm[7f: -h"-MjiCjXo% CGV7jXHX2m' amYhc:c9BF4FqVݖXiy|pÂ^+cVqka:LkKdXd^QH,6@48o7 ῀| с΀ۻf0Z4>$|! wр\9PZ/ދo~fXdƮZ̓QLb dg{l,lNQQؼ VI)˥ײOQ P^䍠LŋEɵ?A涨Opccm~t../MTLD{%/d,.{GƚE1# gLJKS1)EyiyQoQpa5w1M@SX>6Hb=Q+3 >o*"ޓ&sJťω?Wy_l睓mm]4rw?qʱ;Jy|$a;"ySlF\vD)RY]4 ^|`MPon{;v?;h&yk6V-4tE.N :$tbI"Q|Kyr"˂gT[ƂA:o8 RL̊(+iPnȯOw0-6P+fpFUAdRz|~4)~cz[k5;hلrJvD;ɶd~ي #x4kcING*RT#騹/$ũ#JxAgW!zmzyF`(G]CKkOMDئ!ļ[ե#w;z#K oLn慁MC:[fոk09^>&*y|gᵇR)Yo/6#ikzB0*1/D #jYӑtaP>Fw+% yH|fp} qf}m36MLuD~p-)zH1Y@GMf|TkM{!/;zdO6RvEAVN :j4P O^Rbb!k@BOx ),цkBXAgԣIjoH )#=j7Ax7=)F& CsA*J͐7Ų*30 ڋVL$("&dMB/gPĔP>N` Mu0#Pi2@&GW0Κ(MrR%CJS4d#+ڧ_z9dDw d>2գ>dn<ʯ_/~>̗Ì"B-SŚlZ(,[?Ĭv. >y|İf~g2#w'd?\&]ű ->mh'6j(.R`7d}"Y1|˾z ޿fjG`:D\40<ߌȸ|01)lx/Y?s=%|0KYNCg1Gn,?KS`Oo_.V߼,b,4~M^L0+X~}r 5SPplk4 4g珣|}blFױ 3Dcҏ?[! de,<3 ~m#Kf&*o,R]3K"54@R EH  A tt.-Nn*<Go/d:~Jϵl .9w8`xv*ͷONrZf<̞nAd<~{jk̮Y|v{*}zLM&~w=Klڭd4 >_z;X_.|) ͒Q)~,s^|x=~Csx9/|uc1g?hװWL&+ K/ll\gx6M>_u1QNsʞtnE4٧aS|ud_qslz=}{v,Y88J߽_5'$_o xb9/Ȅ&/'f6p ʾPX,sxDpl"ԢjM4O^◗{Gk7 m#)qaf||4X%ARJ'F js ?M#a3uBeHC#=q2.ox'dd2}$Ei,2!;H:%+ޫ'{ Q(c.q!37_}nW:s:a\uVw17^Irʹ8.&aI$u L}/4=Qf&[ >jDo&7&6+n%;jBiB4LX }w#XP 7Mi#X4񦉏`4$ G:"ҭL>X6R)]㿦uQFx!~b $`ڷS8'CQߟ{/6I9=ycJq/RX$TWUnjx]^` o\@a;N =N9B .'XJ&8S:6q- K})֔摮:|oTj7.*"4=2@~rDC))+IQ m8X׈(Lnj"p3cRI{A4D 2%6 Rd1^@ڽ٧ĉpzMo5H)G;^X!,U@*.Ð0!9G,JCyDeDPޡvtm>L܋ IT{|:@2ӧtP[)pe8\wK"Ye:*IK.{鐚  +`}j{I1̱oC}ZǾ|j0|7 QWQB_zLӰfZ`_U2$;C#h)>4`ЈׂC#Hl¾64„[b}]UG.0HXi#mޝSPjfiufSY{I=أD/KI4핼c+ 9.qinD@YL](@u k{uZuzQ:SxׁFzkYՐ؏V?v&.N _  h I&5BX߮b >xZf >֨ PvFCc,%_/,_^Y|yUpҋEJ#)8#LСD4 \cPIyh 17ᔤJ6ȣN4` {!w]g K30Yk yTں8[&$٩ʤ!K aF'JQ+Z !A ͐LP$4QB(]^i[e빡BѻyTԉt&:$$" T]w($og .&ܒdV-%>ܒ;LIvZpvAD[IɭM0p1١vL4-@#AXG%6 w<{S[[LUMWr=5 lvOj G,=\#{{p'T JפaI%4k&k&k&krLI o {M5dXzS-ShPy(7[2+aCv|z^R߭ۆ6mmC߭faE4C[CG5nXblE}5,i{__!P„-fKA=}˵^& )0owcP(f7vErlܹAA6%[$,78hTUt!,˧:a_>ankÒ`|T˧c͒&|h Xخ35 {/ GXOx"WT8!T޶s̆r se6 }ZujDpӪz]~|zi܂4fERA%O[Oƭ݆ENiFqsҚ kFiFjLh{R%HڢM,FqkD J8 0Qʤ I0MSi1602B+(m &J²v*@VO lD٩Mo[dd}Ɛx%8Q1JkRX˔S)*A$D(CCaKUїQ{Wne6_'-Rw"?Kp_5#ۗV9'}L\<_}Ư#4"?Mwk 9?#½u7"\ۻ`Ng%"-xoT_#׿K0Y0 gq-B#$7Q D0~Yv^]#QiO:{%uw>tLґ<5:mM ThXS  #1lj `XbI)&\fÔi)JPH~&淽Id$!, f΍NܴLD8[댜3ȗ ( (^Q,9&2 %`mĩBb, $atL0"Hs$"bx; \S),w[w0 JTR2)LH6Ka&KhE WRmIaa}n}@\$Rmp'?l a=p8 8 {1qsafIZ먼 @[@[U"3mm*cRWGk4\՘@?XleefkA7`檬QABS}]t؛B`E8e$1 9ՠ&`L*iBMRÐՓQo0z˹jmj)ZFQKucCQL3-bE xL:d> >xKxp!ґ]bmm9̑Ou6sч)ؕ"Oӡ̃S` ؉9to@u(*?Mq2DI&Vi6hiGB`Œ `2H02(IBd2نh{S$dV)ƌD! ` ipS $:28EߒKEj]CV>qwkC9]8n̖w`wK֩*1ն՞a$ !T XRpzBdi,U В`!IasĢ4Dш0c&B- v}QsxОfLvtZ_&B',N(lRIi"h0 2%6 Rd15F]k۶+|(b < I{.zķ۵=L)3%[$BdEr-$-RSZ1_Bzd zTC Zș7RmJ$ލ V"MlxՁ׷YVނޭrMTbsͻ1D 4yݚE0qHVE[-̛mJ".Ic tr%_;(x 1 j %Coa.b+7cwv0#G`ofkgqs1*:{ T'kum; |O \@%-%IgIŸ$م$=1T]a5R$wwZϏU \ƒsv?/ x_zO??mݟ`zuv~Rw0dd_Gû g2MG W,\GN1" ߾2L9{3 n{}e%x4lQ5.CӛBy)ԟ9;uYuf**r?) >}ћAf~s(i-&+|n<}p`ﮞ?;ޗLַq G0-O_`^^͛?o[]՛˯_YV!=x<9ec9u}Vf 3ZW_I)5S5t`zSDJ+Xqh ӋpqJ.ŸF]Tm"]PZ tv./󻲘ݎpv jZ5-"+;.,O=_rA` *6Kyuv.cq/4m8(e>MTD1v{Xt=?Z9rhYbr{lA=!Ӟb.ۏ1֮p-S+#2$0 ǂҹ)b| b,y(GHPvRTqEOYG[pN$&6$D;Ν)yim% |# LĄf*qք6Σ8$!FV8b #"TDЎ-X`FYA,n(s U3sidlo2Kvζoa"M+ik Ւh 8ǐRq&4"\V běx:oAY)qdc:"N a <52|!C,H#^N4v7%|LUFcH3% q(4\FFha@WR1 f8V ߗo/ނB&V܊cb%5NqLfllbM$lpS$O/ܫ c FKmq%7,QRX'1G.J925b-P3y^;5f4cvwQP:}{Rv%y'Gekפm5QZ\Pa;GPkˆm)2C׆OZPc%f+`xl%zJõN簐\2xZl68NtiH$>&ʱ뗎c*lVl%{!&脉:b QVY։\ $sSjPͤ!39t nJ Q7Wi!-$])C>T}q`7Ͼ\+W(H_jsgY/ضư_I`_lePmWr{N g>__Ͽ6E6T#>U3Vq}#f!PA[vy0*@Dnr d]6 1 i|d#t n Ll{@16}A'} 14Y(8{VµZh]5ƏfՒTI V)mfS!1*'cBoo9[fO2K[OP;S97CU릴&KR;]!L3)TYɒ57LE['hOϽis`}7iYf(nIN( LrvVotQ(hݫ!ƑSt͌R^8j}e%U/+b3y˘[?) IkpC!;Ζl>ᬆmsFo:/n^("%k$6^GH֘]߈l&' ?=\1][x/2{nֽv}i-Fbcy~P8$/ GBt5`4y $@ ZKCA45 )o2 HxDZ4/(0AzvCQj UO}ZH@4rq"("92]ـKv"HiXj%(Uc\\ -rFh~MNmN]lbGI?y[횵ݐ EgI\*E 7, 4l.Xf^ 핮o\S>gN҉1F%cDq/Yh,Aڹ'o-QyuDYUګ{*帹Bl9x4XjAfyNQɂ 3$ۻX!e, qbSdG]ѫRK5Gl(oY× PHֱ2uFSTFs1^A-D̦{O Qw'Ɇqx䁢b "'$p]a1N'Joo4fTeQ6ZHB*602k;k&;׃`1G_^$7dILV?#EӧR+ >V6Jה)ښ(-_Q1o` 2%{p7~!1T9!fKa !KmaHJD,t,QR&*:9rQUd1gO.ܗ @b- 3A: 9mTX6Ed"4&!ډpFkm% Lbb+2i+#R IBqBD*QbS5QDL#epCS*tsY)qdc:"N a <52阈9b42rTݭqAF.!NCWFZjI4DDcHI MD:M.k126rm"AűRn>űK8cb&`&DRE1A6Le !v7ITr( [*wFQ"p"Aޢ508,K 0subKcɡo î:{r7wFeOD'U1.v7qK37PVO~?'_YqS?zVFc'oeeU9Қ~PƇ{!Xf1 `6qOF߸?{;P@z.h8&A)_J4x|C22&VpEe˘CV``(Ywdr{}VnXuomwoF ;Vݛ7JK֝7Xcru'MnU KdB(KH(R9xc.cPe~ڰaMM*.f<7]9:[] FsT.g>_]%t:c]N]7v4;[]L|s.&X.&.&D.V*VFVN\ FV 4F8 .D0iF2*XIYI8_],b!WKrXŊ.Vgb}`KuYY]%]ZR^(5 p~3_T`s50=FigmOKKXJ k}RL0pZX/\^WsH !ANq Y%: Co'e$R % 2NTBsǡH)(n &O!g!n/ݷR&T&!A h$cKT k`B"rn qn4eǗj]iƫ r4@w1P( xRC:.I尺P&RdK@AA>,t` `w]{!.$p8 ?BV]}pBvm/"颂"iŀ'&+kGB 8Z >!EhU:CBm쌰A  ?G:p}OAeε@ϝ|x(t].Өur܂:]h}Ћ m\v@l\,ja:^<%%'/x`n/Y{II"%CѪ^%ʧ;IZ ͥJnW%CvUuιDշn kmI9~ 9w0Ll&},i$I~)Ţ-{L[LĈedWW}U]U]x-6Z(&p鲑6䬖tg.~elYP:=Z~OUGjP\mj,p'|jlڐԪ_SM0m AUͯG }V"sЌ lOa^~g&hvwkq=TsR?b=V ۜjL֢:TkR?frGչ6:ղ֢:T86DHrEͩ/uB(o}Wr6͊:'؏arw`ë9,L&gI,CvV>ao7 |6>k̃z*|>ĪWՅ6/0./iy0EmDAIzy{ q3a\);GOq[25QNk!I?]n"|-M:m QX̺nh@3hjqC]+XnѺ tRXǺod*|" [yMt>50ay:meNeJ[1M[ yMvjancJrŘs:Hw1.ܦ&hŘ5>.ܮ&0B/ƬO]1 0_YKŘs>3G.Ř[&GcŘscH(Řs 73Gzs1sʆ1w1'5A7Ĕ>AYn>ތ0tBD83g 7f JB23iX(rD;pmʫ鍙t)dz 7<ʕ#m ?Ob'd_p.ݣ9`?ӫ~Q׳W04y\挢\a;SroNTU?Zto2zݫh~x0j3جzֻߧzzG$V6I^.gZƅ^0<^NĂ`s81>0/ LOA)sAigrB"øV:/ J9Ɣ4) N@Rxw@c%98pfG׳a1W^0͕ƁU"ʙ.`^/N OԸVfx`хʉ@W0D274ZNB+К-l>Wܠ )ك,ax2`Z}2z3iWz0tc)Y-}K`xXbEuqwgh J^3IIn ߾nkK`eug=Η2+H e} ? Yy!?30۪dgF s6h‰"Y>]cO0űJ`L47>>RKIDAMۂ~5!qFFq"\H~0p&CщVJx~tfqrz4Oa9pb ŗYTK({=ʦf˷;sЌ 8I:]ʤg&h}$a*K F%_uoɻ&$?:_ŁJ!X`_̸!i4 5+FX7) ]8^]x`ܐ`#Qu^eS1{u38 =:rt%Yoa/F/{2#?e _ύB6L*z]M~c^TK3LWi)HS>Ɍ B纐>He4_\}=Ȩ՟csM^߶M[GuTʍs0D/E1xJ0F <i(kI[ND*vmDr'{JX=(_k);.y*^LzDlIyת޲cgd^1egt 3O VKk6Ӣ3;k(q.ͩ!lH+%WTGahi I@bG]ESQxkpRLV)GPp $p zZ\IZ!(0NA[Imč{DaT 9i@" {:.( kO]s$ع<̿a N8 %|Ւ15dՂnYsڥ`&}j>yFKmd9'[cf.XS.B0NɯLG fVɼwbׂ 9WIpxZ'?_~e LĘm C{B|?-Wڕh mKt1Q9ٹnݪ̔pJŀ<.'j4̢͚7 M} )22!U2#l9d/YYf3q:%KZzuY[2 DA1Q.Ǵ8 kҼ,r[}#C3].M-%.sL&0k\-~U.[;_n9DGBI28@:,`^Y]U4F>GD\NLQ\@O"k][^eTg*8= t@ A>6X\AՔY/&2ћwm36g;-!(&"wGo9o--\{Gr:W< 0$65q'-R^x,# =p p<0Sf=XCLjY|%~RNE j@j-RHS8S)yъ!XoT!)aC3Sh!ڠ9A(U1D`]ς`[2Ih 'lX;,!.7 I]198V0q鸵3%I+c0(rE>)5H)Va 2}l#})K1\+VV׍gRBFuHߕf+ ;YJ{ͬAKmuonH}wfirx.j @h.RX { m%Q.A16fFN`hw{`)')ܐ;<bc6<`m1^F!4i-I} 0Vpbu('}7}Jj\m绖'bUi㐍dH߯)X-m4p#Th\nHbue㶼gvd%d۔U7mAh1)ڐvI^5P#kZh|'OJf(Y-J55IV0$s*oAQ#=dfH+Ʒ?I@FYOfYM'2"? CsAƺ@HE Xxg3R{y4Hn g H'"]! B.A 4,d쳗1H!ZPX(pX @n ɽKe,ߔqht46!Mܡ,H#=#LPqَK^z[%#Qxx+n»`IT;hDܵV`Y]ԄO!*#!MTg'KzR6gZIcBk?ut&'~S,X GWy{x*Oզ"{y6*PlnޕqdBepv=3/$';Qʼn( jjil՛l%A,::[:/V)tg^u#0hqxQ3<7ޯ9S 9)"/;XQ=8ܤߵj?i|oeϵВ?_O]$XKX_g^ij*sm]#AyfաN;$s{sƞ)",)",0p\{0:^6pR O Vk"hd]lKC!6juz?EU+Cl Uj5_mde]s>3痾{/ ymoJ;NIc0{oݑ:F[q2<׻ ;5/ >O1"X֦PVK`G8?)LrĦgolrA+>lz `h*͵N*UFo؜󆗲WOB# 4b1I ,(4&(.jRTwS6kpcxR-T ?}[rn˳dJ;&,L0j03f ey3,^r!dey7Ebܨ{B;$?k!hڠEE7z_7B0Ǽ28f n±a5o*G鵋2؂Y%Wo`inAȀBG+FsQ } / }&+6}wZY n ORΓMk݊9 3gķ^CSze]7W03W>qڹ4sW)+sJރ9f5Rj7+O)'C4ƣ*/M.?w+d %hѬ%{^GnC rsֲװb`pN+LFPԉ +˽ci9Br(2eE|Yu5mnRwo%d OUD(-Q6F #F)ɴ\_F`T8HL1_A}AQݦiuqn:Sy|7w۱.>lInJDZe#$@\4*@D b[4@ʉ8*PS`h;~:|.ƃGnAvEG%bF%KD~6#[9OàloxP[rj 6RHuрrf>ưȣT%[ras`[ԁ* ,6o}Sj9V1-dBLf J(c(-؆FЄn53Ǖ\6"EdQiLiW4xXjo `sp4P\נYT^],N:q՛hyZףgjTч~yQ5Sfwۛj6ߤ^?&yzqt@sE9r}mtur< fVտsb./ϖ9M 4]sK!o=;=+, s V??􆏻*1M9{**p$1%5!ɍ10AQha4=t ű/ ݢʺ(ywZ7(=_@\4>}?s{M:o^0]J鉽}Lz3N1-{Fc &H Hkl(_i58Q[R֓|5\6|q]_2‘z')u5&(+):VakpBL_!MpZb-G e\yQ1VtŔ)$0)c4҂ L^m'E,&K֫dV롴|zO4&Te.dl}+a`)8O<%bΐ^Ks Ns ,m\$x\\{xyrvGQ3s./&o{#o姐6n7'm,OVav~Z-g:8^/_DB/>TRup8Ks2[igoX,W!^]Vܞߏ;bOy>:a N=Mbs1&{Ŝr6zj3% |d߇˫@!yĂByx ޜsbv8~N끳*}Ҽ5^2'Kկ.˺% Xg^z'tADۉ x'tAwL|^wLjyB'KntC|=LjyBkegUyz't@!II-N< *7mUc^G\]BBaI؆`9ՍEStucuNɋhnEȷʐaȌ@p$T 5Iܩ)DZC-]_֊f@KEl2c7b1؍fMyt;@vScXa]j L|{T@V/SL4r6C ^L^ƎZ,-~.ewԌ™.CB Er6$MIl>ޚ^Z?w] TuqnQoJ'CVaW秳sYwOtŏ/ĭ T{wgX-)uu~_nWتϧ4FIδ$ˑyE[,\oP: yM+hݑzP`Q4 =5(uwPMc CL>Z u'y2/+EQOHhpl,51bwAAA*1 /(TZοhDfDkK,pQa,8rXݬ" nVtZ8I{APݜƙtS's=&&jr+j:[ -SV0:^pY/*; wHM ~I|8DޭV KZp*5cZbzYWC|,կy~;͙|{Q+kXދ7u"ExEZFM1DRYNPpEְmsb鋊b+&(9ߕr_j;Ҕ _ܶQzn}Q*6@awQ &\ jlž``"ký\k8frV*a'rzLl%S?X(sdV/RXEkիKb`/J44j+K݁'{'$DN (SL ;ڋP@mt72bS;1>CnRCCQ*س>pHh|>TOaJ53Bk8p&HSH\nuZ~M'.mm&˔쾙teO&~wO>wTƑc٤!όTsjb_u'QlI%ldW=4'CS=OO"+ w|%pQ.CEP"E`eWgܟHeE(!y#4z׷kwʜ.8tByߠ& 8VῨ(5LoW{cgV|8ھ/SZ sVȋ^kI=e>\iIs<3 0Uy;njZ5r4\{CPyC$(t1 ]rrLX+D9:4hA9!_Vo|Mk)FM^v0Dk[Fs}x uv/r&XcϜi`DlFʾ"S$,R\xDTXAcsBFhjps'I3w}ђU_Ҟ)iu0J <{0D0:0-aWA;c@{rQ\^4{? Kp k(_у")Bش:3VI;& s$9XX* G/,5AygUnA"KrPx,Cl>N`h*^ 6;R t8H SèF1pSlyK]`0cjʐ}u0H!6N>k[e6u O$*n;^F@Jwxe꾌flhw}g|e_K,ݦdȨ` E@>|yK6/J|MܽiDY_ ~ ]q9{;Y{] /=|{p/`Ҡ!B`bM*c0k/^Qň)'v= +ҽؐ Eq,Op |CJ N05&ݱm%㨫%?*x08}݁dn "0 `Zأib!F),yx6*]ӉAyh3ftf 0eϰb`b~Ȓ"`1-3=T |=>``-4K & <)a Z< x6ʸ_w,3XҌW\(26qEkvKqu0 %)<]*Ym_~9*5mوNe&zCq|HqpxR#..0'}/RXIU'#.r+c͘bREqhz ̩{<pd\x5 !X Dv3*?|"g=hEYž~<#7fEKVl=|9Y#]g].g߾\:&=-˱ WE3B7Ĭ2b:RROV<u|sZ љlaXx'"x]Ps ؝5?AP3"LR$q/Q s"mh3`?tTiC][MOc2T A:zbx30c2#ݍq`(xh5_eބi|2}4pj0g~'R'7S}|">+6 ͢UttqX]{5?ėJjn1tq9c3,`{?ryB< tut8S<3<ef#jB|ڸcl&5;\D.jITu9 yPcϨqȘxȭFj &5{삓띾>PqJJ. ֝3O o:Tj3lpK'O8X2u+08*:d`a2rQ% 8eF7kY*R϶"Ru[ӫajWM+|;>GO60*dGzg3)xz \QN6^v t䮝tlJPҠK$Y .MC*ύ G$ wTrT Fe[ r̓8v*t.I:*P? zHv#(-1VBX Ka\RK0 4ǚY$Q&,w*l5ke$34;}4A`NW#3%||0Bt̾%tS>,#%^ز*(śU N %N[E7>R):3b2.8QG̢6@^5޲=g55ZGr?m:Vi?+F>=K_"LugWcrIڝo8xkYăL0!rJL۟>yjvLtVw3ps{>-$l_*R_:P-":P(SDMMyf98}V~-XŸgwCg;jaз/e(gXx$ Eq;I 麠<JFM>ݱS;Q_+.k93tZ[vano CZ1Lzwo'S}xc*OQLսpQ{H1Jqr,K6;&|Rנ&w0 K::t>OΚY=?gC (L\|S Ukv-};5o3%EMk9\bis+d.5&+k 1$2'+$LLY v/P&CY%xGL;lА1b1捸(k@ Y/%3L0d"D0W`)N5? 1Sű"8!' IOGk61 hJ>HF&H'&[5--Job6v#h q?>^eZvӐ-zr&>+;rBjUg&-1%SL2˭*c:!12>KyԘU{m3&de#nj# h;gKj1"%ؕҟ>7H܅\g,w"S.,HC(G4;Ǯ}LJ.wJwI.wj<q?O>8누\d)&ep qc VJIl5EJZv):!C &֚--{y?8>, a0%锃Ϳq"JYck1v s}4 WY.qS .1wiʣ ӏJTHC&ˉ~\ \nDZBUz{OJEQra:`l=P0/ow /4Nh.E-\oc8 f x2/P?'!2.dw/l/*8cChS zTIlD>a|(Vg`TGh[ffc6;=,4Xb1ۺ\w1auMqV~D,g}ܿ&Y0gp9U,U^l:p<Rh"% y_ L'nX:xVt)1 9ӬboUPH/riP*KAq!zc#^1cG ~~AJ Tw*D}TBR6xbH$ϗF:L~雒;\gӹ,r:DZBcU\#}<\Tr.]=N$Q%tGs2MJmy[\\WZ<9aY!+t<Mnx~ti-f䲴p*2sME)j=1y뮀Rz௫x"8Y`G{Ulo]_X x;'g0 ; GQ|;Gw݀osύ748#%[HAa8+ PIbnTp\Ve~*Aݎ.r>``V7pjZ}סz0;! wz|!%WI :uǯ*BYJVU#MHJAuzCtΞG{8sL9ct)! s+XgMP1fVA(B5/n4 IPsFݽY>mr`힂 I&Yv˪km춞8(r^[ʎ;~^&0SYR Q9 *eZS-{`sV:?x\x3 X7V"(Tr Ժ(5O u2!!M4[=߯:H@#`\瘢)글B!GFPR 2+ԬIYVD>tZEgEPHM/R@lXn 5#E)*%KgE%ѭ# C=d Uv9ê"iSj,JIei1FRZQI PnXMC%RTd5GVŢ˽Ѷ43lmՠ3B9ic̵FWL9hM︕;g9bAƨx1D9&< L^sD6fZ渽?pnZRC3#eҠst&,'Хsm +bT:'@cQZ#fc'+ErOkh>7ǔW5c-D+\BA4F c 5C5HP[yA{S `6[6ljPC6u50;\ >(#el B,2$>XaPA& 6$~jjIV XsIME++$R 0$s4"d\2! jKrAKFo|IPTj5drPͤBT%GGKY=.}iV՟dLm91,p{hzmbLk» e7SrI~G< !xQ-{GEbyR2A WNjf)0 ewYYX|6z};4gvvc;OoƷG#zM͝6hkOrs}柯у'|k4?Lbo0{^ !9z p6{ڎzS[}<6rIL ×6 R}%dfxW,oV{z?[ {4A0R D4~\KA39%s6Erd#Ĝ2c-S5v}bI٧I~/i0nvof~={@ SsYq%ޒ$OqQy(G uF5XᔉT^aHq_Lwe|ѫVS:(g`fm8hU%TS3jl1Ϳ՞X4WRL\g,R]yRjů!zjE+87%؋V a[^AT È&ЊiA9[eL`RJ9H^eL-A,3_pbkrJZK:+L:RA mt&jg KNbO`r@GԌG?5puI-"]%*c'nQh ȣ.'"@'-,?A)0IZje%"\j Nʵs٥1rۣ\C˜]Ot~-&M7׉ՋCLAMP> >|m \evvxz ;]+兩wbz{ \Oz` qąc BU2`ttSǣ]}Qǥ&Hk_ҙa^Q0{ѳG '\q%z:\4RIul$C" Vuym?݅h{0(T [GL8oœvnusv»γZcG0\0I6\jj'ޅe'H2,1ARÙ5 DYl9$4A6;`WN"1y0 @0QF&|mgq!`ɚNyXr8$avj)Yt2zO@sRXQ}K}DřtgWOvrep`/H327[ /`fĎ ӭk@/'SVTpۛ.iՠu(Wح8:mЦ6WvM*@݁ܪFfuzy5(]-x֐K{ŨsU_.'s O&JwvF;u~3 ݖF m ;5=C4Nf/}B%u4Ƕf4C? {?ޮ|޻-7w;z'[nrdm&[ W X1=7Bz[ OF ~OS8AΩQ)Ux3<LHCKIψ\yd\ Y!Ox TJQERM>̚l!pOˏCOe H!Q:h\t 14GG Ғ,IYkCH:dqE DD4xZ1N I%?%w?GZSkh>JAaO:g!L\u׏d#CjIvhc+|px#J2gBF5~q0╣HIRLD(zei9Qu֌ C:7z#j Vkj/?{-E8\'w~ց-u.z35sQ@7nޝmoA.ᶖyg[Z|8Ѹ>'NTtpbf!kٖaM_矗JouM[3IΈDYřqw΢{%€8Q%f3e+~y?Lu͙!aF Y}_E(,wli|9vgWZVS` s$܅aT k|.yf>$%8D+&(በ~7Awk׽b4 9Pj^ڥW=RR% v~w}_TwٗGNAUZ-!7!䊉s bԭX<[۲j`-Z#=pr%y^;cIX=C%|4(:MJˣӫxNJDCݢgIwt| ??S2nӃYh.wٶc /]_r~};Z-E5˛xv4;W~{?ޠR9+as*si;K)GĿ+HP9@VeZT\?l,pɗ2TXXsXE*iUL䝔F2OZBKy%~Sj^2vVKt2O1&x'k-&o4Ǜ|.xBzoжJqc%pb9bGFjrK A2D 00NApAL쟵Zf,TjT,vɨ [:$h!tc2BdG0^(-vQ,vD3vU;X%!\R0-^ec`L5 ZyQPv@'k;@ O7ҹ[8y#Inw-j5-j3YJI%nVoq]NŨ Hm0–TTuQ =mﳝKfWHIQX'jx?B+q".-!mۿϏ|23ef23G;IA'BIRQ)2wETs1(2A/rN8ɵJX}۾7*'X9YO[ӫ^1:~mfFҚ,Nr_Yaɍ/2?F]Ө|uO.˧QeG#';BD(KhbM ?ǟLry-uj3xb/?r7֥k*˽u z>-F2EװknkA-n [=CjaE|G V/0Tfy:r-)HD &F'rx+dIVEK^E?;/6YMc>yhpmt)rFaQޒd\in\r([ڦ:cK0ƙ+u2/BǹD ·Ƭ$^x/8PA?}s\/,޼~?[TLbWz 4E?<Be~Uta4B7wg ܻ3⫆@{#{?}ltprQVf Ԡ_noqru0p@qI3omAVʥ8r ~B}Ə;X53R$_(R s&] n$R\k 7JCrR]ōp)8SMQPd Ъq$UBo̍V*HXFQ3N|%9GP}Dߋ%=)]Ҥf̵tFZ8/xN2=yF%:KԌr'1HqБKhkjRgF4q9P"w%KПV'Irm3FL^Rz#A8$@C! $%k #SBRZ{|>4Hia lTkҳS ⒓K b|>~&,Ly)^^mc*)T;??^W"[?}0J8rQ~$x~_2"vxh;+ъ#DJHY-kD%(/4j1Qj@`vbowԟ\թ}CT ;~6k϶YTAJ+48*zDzkWT"{hS4+Rv(rGP>+~pF39NwLq3#adc*-yBڝ=k5u5]KzqQj洁%kt{dd*4LN 5HȈ2$X* `BYq1p1CV}HKNXK"_J? 3JJ˚ޯ^ !+;{}yyRJ-xn(&zL%^<@xkZ' pƨM0 gVYq,،zf[%ؠv`51[y -J'#SfSed^ ^cF>Q){_n(ŠbS<.ƬXyMFNusPRSO'`r$G-zqYirN:`B~s"*<zR^?P rW'FÔy"oÃm2A~R:n[$A}0SO\T6$GK?` t=>xzqĐVCnM-AI!q2/듿YŅ;:ʡ+ΐ~ r,gPӺtbQW7}maϫN? 7O>1gi.3RJ&a6Hz-@3R'h oFY1JnR&s&U\$O.6%YMlwl}0Z;ÿhy_ҧgTs ;Hֺ!Iss`bdIuv(?J,iw%rK7ΘNoy2MEW$pOUbis 0H' 炝xZOB VxJ`pY$\SenzzĜ•4Cیy e`53d_ tr-OZ)1,i]qSS ܉˓E keP9ӂl}< %y"*uؗ.mWc ئ^T M4,)r,$$$͢p#`ź*u՗X3$+nt8RbX*\hEE0p9FB6 Asb `F;OA"z\K1sjEjMI*@-|lO<ROQ-:/>3o ϔ :`ͤ& 1[qlޠq3hRk5hIf.:Rv9#(ԏ J -hgQ+i|<“EfWŢ\qƑr+Q`-C^[2XIa0%["(B*?{;Xt[{LnJ:dM=D*B`fZe MtLAwT&#:?RE+tDD'`]~D Xs֔/_ ravYZ,04`IOOP&J<טjɔDq5R?֧#y33]4lenP"9O#@;E"_莤$wxVq+ M).D88(ּ(,4ւ8ΫWJ{et 3k)9U0!a|E;}j ` Bvm?j'+* 5O!*)l@܃Cd!8 Nrƙv̩"( JYKA:pgefgɽ7} Z#Ao N"B)èמ[D30xܰh t08OJe2tIJAJ+3{H(x[\Ď{5+0FH NJxW=bMT`AH!6X 9qM"fbzI7SS9h֋/SQ',66 ]E.bf5uWss'wЇhۃ̧3=߸*dk"U;/(8E$! 2:iHr!9*ZT$ӿCQ؛eR`-LUThIiT EJGx  F9cϭm^ag$ ߛ`eYDg`LVFzΐgHNlVO]g?d W{>a,lN#٥8{O^N뺇{Ij^Z %)9_l`W̽\')/vz窡~5ndM1f Bd7Ig%j5զey(FBV oD"7Hij"eFDQ; %3P*;_X  E!0v:T ᗪI*cw%~3睷1 ^J{8d]ԨQ32%" \=M쉒`,(y e]y,B|jk+!x+[\R5T}فKN7 (k6!UKPn-04Mpe%#yc'a3 =(PDJ.r"B|SFScG#h\$ L4 xd*xÔ0煓Z=#lAU[iaɒHˆ˓Ijjs7[JW5RfB~ s=7sar֕ WA^{q ]Ko AeA!>S6;3? A _.QwѯGQ۫R_qә}位u3pD=|'xLZ>^ȳ^; G'ae. v~o] %/]3G>(> knjRaE-g¥GWm ߨWX+8:85oM]b"f.4jچIc Oly꤇?F93ʳ"F]wsMKeR/c+՘ LIvx &H%.8|6[ם[Љk') 4gAJVŸm%~¤Mw߮~ats]ETğr6Gn^p94j;QaxAs'pSkzvKuܕڣ˕аȜi^%S5o!{:"o]l|$pSIۻ[bE2f,I{Ʋ|IW//BM(8IBxsß?c?{vˆ"i|+z>گuUxد4SQ\+OH^dx=5PB߅r-|4ÅT7ݷs,$$$͢: Z#c1ZGnM4Z"3I/zYyzO/(`X9;ɻWo_Xg% c½r~ wv1:αb]붏ߝ^_}}y|7wyw_:yw|Jo/ONϏ_? hʿ z<9~w [/rao~ ag+~v XAuW"AAk>we@9?OUwrӸ O$ nK'0οNOcbؽ :9'.sޤ4zBҗE`1˔UlCy >;W8o`rw 4_uvqT Fn\-木s?/(Xt~R!ңˣ!Lū Rƕ_Wؗ*䡀y!PϙR=X *69mT%򁔉"Q蝯}!$L;Yz6z8:Iy! *OqP7YcgDKU;Ym+׺jmgE= mP޿y=f ät;7AS7UY3U N+¾LMӟ &<2?[ ˢ܀2]}{2s@붔X)0s60wSmT LT LT LT LT LT LT LT L0FIl` B6" ԤK upб BBy ssӦ.rQyb-M%ލz7,럊ӌA“S.uUz6h0_NƇ_4g`"; zPՉ,Xhl.'}gƳ?`e1&#&(`LReޢN@[[SefP&"fNXQ4Z"GJ-Ă~Ua@2͒,tBj"+ f=^hCB3j03j03j03j03j03j03j03j0}Ō6Lu0u5Ѣ QByH `F3bppH6&LrO3b41 kxa _f{4CWYp BEAAAAAAAA-0h-^R ,?qIn ndkpᦲժ_9# _P՛I֤_NTcH~B$?pDr> +چITj#@ehͰZO IdJ " U"]3JVpeF!ќ9K$EԵDgc?ˍ֛rE-'z|ġ+ϛ'3lFBt q$z7[A7YQ=򆽆=_c?-ɂʂ7sYǝvM&RR\K!Wg߂k|pYwXRԭ]S_qrqZQ#J=q 6e:bpzc LeA4U$̨B<)3HQ$6uZ)02`T_vG2X&MljUf0שfZb9qBNKv#!(1)rK0*@_SmS +g6[>Չ~j 8Po/{Kd@2)I,vS--T\B9F # C(aa&L";q1b݌T}ʨv{0r-Rgfi˧#;@=NoNϦ]CFi& 99o ;Lt򛯻>Kzq3\Dád7 z)Xp4.eJQmWXR1;8 J+)y+%eb!̕/ǹjCȮA)~[מ5eQ$ִaT1-ٻQAXV"P?EJ` j[% &K{t͋G >B)_{ zR>&n/O" %}>MXؾOgI߶XNIwA߭R1 lCۢ{/ # i6яW]ls.@A:^uZK{]ܿI=9چٮ[7 qlP4G<9[1O٨͋'{0!adXm s3Kh/#XN>93r YoXc/[섊J-Uqm+S9%qc\ݿw3qA\(/:g` ƒ~1HT *ȧ—!7"o[ALiyv̷,K^N9w-xVkolֽ 9N3v**)C9y' hcf.o.Qz#DλQcj E)AȞ_2p q?CrlHj#& ͠M$ Dp\*%X1\T;Ҝ` ˸Ö)*L\8'9^Myysxj&>֐='⇞udP=*3uMr;`7/IQp k!2"8Df6Vؐ{& )&^]S }Ü͌uH&Z_/) ؁H16L07r:眻>¯!}T0#HLAoM>L{Dwhj1c=zㅡh ,`lOr{7b f]u @RLB;w`~~}P 鵬=ˬO^4EN畤ڣ^׸8He wb=s^Ͻ6Wl XseD%ܘe h$3$V;LD3l UB|%-q[3%Zm>ic(̡ $4fB&@֤Z$c4km~ʹ`Z ːg=X,6!y?-Qc/lKE?~oz)w2(`Q[Vk#D8L݈diy4}zQXeL<43l8IhI¢,il|;/+qHgฦeHls`yFFɈ*=l%!Hl4 A6^p~[&U|ciS܍7|i^GϽ{=h5DҔZh]b,0:ce\i,V5Ҕ(d)J9|I;pBC,2)b$r$Nq1q"o-IƊ>Η'ONO4,%% ?\;e}S^F+W[Dql? ="R󚃢|ŪLhFuaZa꫘QĚ&H&4 }u弬x]!mu4(6FH"42NF c9jidl4WD> 8y\:RMY%f N/?⌹0gVM8鏩c6`xcBSs૧)hdkƓÒ[}?Hk)7}Vi<76D}a^qZKT[2xP`bi(Fn#)S5](F!L6:Orv1lo2?uܯ~?yuf~@'2%R.$2ц3(ӄ,ԷO8e:8SZoֺ0U69o`:o@ rMqB_.p%\9ݰե3֋xSU;ѳ~hG/zw5\i=~_^RPArXWdrh3֓N"cݳޔ/Cx7t+n{l0|11;7aL=6<ڌvkK{!A_DI 䕸 _͘)x7c> ͜<PC2.J]Yo#9+FmG=vgy$3eeY-Gy0}:lILQ݉ntW"8>I:5 'Iq(#%C6n,BtH\/VQӉ[z [觋3\`ݏ<J+tc3 ^^|!uAÇz8Ueg Ki!0V5MZ "H Rt#`{砞\䟈C~,SFsW7B4# N$*-C 0޶R1̥]Ud>/3o;g`(/fBJW:LiFrm08j[1 crdYN]l0Q^e9}G--B7Y,C,!2?*\ wAW?xӮj~ӧjz5ګWB3WXDHxt*L~m J9ov? 2V13AXd(FL'Da.aiRp[-? x) }J;KT Nշv[#4٠&'9/\y 6cʫg@[MԿַKLf`| C-~o (  j /(*I tJs6WkGV~zi)e92) jm`UgCOjd xUԑ:4V ΕAk(H$[t`G(bgb4'v=[Њ+OOz c6\UvQQ3,sU0VT! )ԍW{M[?zƹ>Յx&t%\F\4 ŗs&o dԁ&j Y:oҴ4B>Zs%Q-U @JVX,gqU6v-Fbʹb;'Q ~?GMȋurâ`QSR[ΘR2pʼNk[Qjn*C8&^a#PrFZtGVW*Lt Kam;gRٓKFI83J78{f~y'h]ȭ}csƕ?j@+;af4.Xk@naV*kjM'!WuhTcLdSZWWV{Xp,'|۬Fee"*+ò 9͔C+>#4&ۑg =]]~,'_}{U?wѯǣ~xWj}/.Dӵ8^uDXtqƉZtBTK~ j'okjy ܍#ÂsXç˶)Z*5\)$S gN!Q̌PE%QG՛t0c.L cW15SZ[ 5Iì\BJQ+~Z_'>W=z>|"T]b|e'{M.^"mt @"C~G qrȲ32yv;Oh3˾#&\bp!GZp׻zv?/2J@ , oI*Ԉ (h4pk[?ZYr,+$5luhZ?xxy1\z>^. VhePE'U t 볳|ݨ-Qy{1xnŊ ^+T N F0X$#|E!B+ċi Sĺ)Pn%yEH"Kd]1D[Mh}\l5;4K'2#/:E-L~YER(ssR g{B{m|K"rOr6bcg$!6EhkvW:$ sm_㹭 T [*<2KF;Xw?Zo1x/hSpȫ#n7IhǬѣUJaT"I$/y:ZcͬG--A}z뺚Rm5ouJeXr^WReNr=L̋. pq3 /ihLsPP4԰w~% 7PAr05PVN쏡.8Y#(ڈ3M H`yDaJ~pL7~R-ƗN%?⛑oU*;E1*٘Z)fߩJ_:^-.f]1!kh:/`3l ;dӂ݅49⵭N#/PM5ڱB2A\8@>c!͍.^5axUAyK%RvF7b-3YnFaiR 3xiUR#SQLˢX5X Xj!>L5U0qXdFr͚gc1[8pY;2@|h)!1C•Qc=5WNaTz^|A[=2.fؕ r n/QhX Z/cc-bA`y& ,kvKp\A +v>C`+K&FVΗY_)dڱKvL-ʤ* 6g|5`AiWFVr5X!jSf;j%m@E.%¼T\)1P"+}B\8@3naNT (pSgW.تLJpc9S77(mYg745*%g\Þ3{')I.;v9ld$ވOtݏA=_}"b62DWK0ǃ-B>K >%-+ʋ /ոˋ܁̧ybk%㢶f1yL+s+| pog2Aso×%Wr(+)f4`p7<3Pc5x-cp.@,i:>"{EXi31gB3d+p8[K"[R=)nځyzN:S@cXg6z_K!Lusj{ :g@_$}Fx#)y'BF2{te2Sa|/ɨr/p5yO)! 9Ә q4T;V)$Y5x 1, %0ъ~qA6peIRA˕A1lX6x';Ҥ88-z`!6Fump-g[erغc?t?錩9c_us3vx6A}V+ (ZWvh?DleՐTy늣$glfCuHRH"DR/5khk#6qܝagƳ.E8Q5k9LxY7Z@dMFjogp71p͊:`nr|sPa?.rYO'q%PgEmYzI>@%\S6f{Q66΃vrFQJOܒ3G^?jԤ>3Xz9uDKEn*SVypq#,Vhkx-cp7pif׮CPk a*4288;*1.@f}H05+dg9?[1/o$\69WΉiyFwMmunPܺĐ\ q]y%)۳`SjS9=ݿH9H)p7ddUJH]-h%L L05?w7mg*ۺ֣E1x.~pN8SDkAqucөe1'qL Qa";Qsۨyf;_*IϘi6k=y1`Zʋ1[,H-EBZzh@ţNH9'a\WxEZKiR(G ƫ_GtڰTqO֡ CUM%9_kã/뜀#S\AL|vv~>q, V4QVlųaš>9s>VaW+Z3Hg>'F z2T̬!4&_6OρV|oyJx%~twW;ys%?1=S^yQV5X6U>۔0!#_^yP-V5|[.|ۺڗV-UE-#J_Ieˈfyt 5O{ a!pqF^f)ZgZZ4? o?GM2F(-.([7, Adk>lQ47 `\%"vy5ڇe ѫ $ScGC#1&2PAݞkEq%60^j»K,Nim=,@B+ PP^d9TݰɇID9GTzO.* #WUj'J}E뉒wֵB< |q$K0hXJ,3^Sgҗs|gp47K_W%IV=g5DxQ]bo2J}L O\ Aȇ+B۸\G`?5~@^ 6T#qvWDAafO@[/dph8UHZRX td72:i~g20.TC`Zi׹#n5,RaQh|H}D?#ivi Fن 6)yk)Nm6_|3 ͖9̃ĊE2yFU@QJo0htdK*g_'UP`oV\yN=O#fhfFT4-}7ơGIpcTGʅd\@( inάlC4N xH(0Gִw YmL6H֊%dg"Y"WQOFoͿLrBCJhȲ`:  ό}Э?7RDͧ.;^"]^3(vOH'ouO,韏əfkcowIhR4kw./xfwևw[g}~swj}?Mg|-4zzkw~P _Igi޴sw|ls8oxȗz76kpBĤ.ootg/'apKCx|[3D(F9f^wL(I3ti=ylCqku%`n[s? ;fx|uj@Q)pxUzb}O_<1 /OE˓N}'e;I/GY<8O.|>"c$sjF.? =w#{.打G9_dRüvw^zMg3t̷4MANgw?9;{a4aងN]O1ݜ ӯ}Bc{|[mr /ko1ooG\ZI`ȇAHFw{'g+'_cc ׋i}?޽o?~$K[oݓ,c~j1$]>tWxNҍN34R~Na\`z_/b]1g77~ aȎ{}Mǿ:Z?кFk͐4ԥZMxr<7LpyntғrBkn>@tuW-jG20Rm0M3Ől65hkZ-% ؏!~ ŁW@2rGQJ8re7p41w?Sܑ5Z?ݱrhMz۟ۛ?k%|i%@~vw6BT6Jt9+Hd]ha ! "JY*|.e)hN]__pӊqv˂RX‼ļRkd MyR0 )JZłySSc!0>Rp4^!)!9҂.ȨhT Ϥ7GlC J60Wk^MEv.LCVּ$ܙz Hgo%,ELfN!cAڱS)QR8L:Y:S3/Z:@gusyT뜥C:=XD.tG*!,8O@Hhc9x cЁPo &h}PFclyOTנ9zvkogӋc:~76ʏrol=SPZSoeZn-W&b9#|ˡDď+ƴ aZrӒ1mҎ(4߶g|?Ғ)%تQm^zډAnOzhd͛yO)r^#Mc>QɊQ%Zn%c|\"q!S ͋/c}R ]ptxJQQIJ6/ՀaBf{e" TpOb( 85-pl9ltј.!ySx`$ (% F0by"Yfщ9kdr/ӖyIfA>a-: PC؛;mVm˭"FH""TOpiC'8 z%,1p܄oYuOn %LooooVDJ`Xq+,]8֕И;|R]$)̷0bǘ8bEwKM4S|VS-tTӴXbr`4Tqpe*.mr_ DJDv"کjJ *%K͎;%t-8-7[)qU;lbF|m{s6)r[Y*Xnkm}-?1U$#yuTf_9  %sj?~\{u&#贆 JtbKhM3"=\S爒N1#BVZ8.I kR+ /EGh]J%9xA<̽%_%BZ;-([f2R2o:tA1XYJ`8G .@ kJ[`E`rLr&SM)[Uvz)]τM.T*Pi@$@#i2K\᱔wS4XS D"B t8Fܹ DBe<ŀh^-x]*Mˉ6 zaVA ٴ|4034Q8$MR`1ʤxZ0s)VJQa^ZX0X!oʼgS}'߀oumk[1ڝ#4K xuAaK1/(H5_8Z _6^Sza>"%8pk5W<(Ϧm/xW٥}۾w{ھw{w(v(}~? ~w{9L$JSJ9>eظOFcH k'*R)bpSsЬ=4B.e+$ Ox";Cxᾆt_ 氽8O ?\'xu@upVШp#VSF4GsTUi*G¦:D\%BgM8^tO.UOps8Zb o ~?67lSuKg"Z0ѽ%,ꔎ9USUiΛ AˬNyd&=vwkWkSA<nj;kO.+!'ڐ|o!g!ܞ䏾ʧ~s_+o|cʛ&؁ṋXVֺ[ܕn]>Z/ z Q9߹HM&RIZ=aKm0?$G狛?Cc=f'^w6>ʟHp0o\9ʦFha+M^-I:6+!NYk{y@|%W]YkѾ2dxxevIuﳲ~߯qOot96~m.I/Lw0}no\zLͶ=l6o^=WfӪ 6=ܷMhP/Iu3F#Q:#c͇KKKBJjm]Ta<쯞,BV ves=?4}x0M.ع 7~4gC&,]7U߲tc1?KV# tLgHS%K-uNn%Nȍyֽ܁xFd͝p+E4c]E I|>n< I5mwϘ;ᎇ}^跻orA܎]7w'=aHX[)J7FVD/%E9M.ZRە桿aym# 7k1ERkU(|IQEVm[S4ILV)zVI'N2}L8d7ԲҀ8ضc` =cr:>}>)Io,jeӭBieSZ7\mΜsA+/b #U)\*>cQEz*:})$Tڗma&#jԯesOnvzӫu|~om.W+:NvTI>}jp#pE [G1$JsPځãf'mϐ^o)[_; 6Bs L+V\ԁ LB}%xSY!3U3ZkSU:آ2B ]zbU8`ۭ:~M,Xwth"Ձ\#I:A`|ȒDG&8XP,8n!+gK{MdL|bloa\AF Ey:X0WRLG.]fP1SvmxjUFbz'[(Ĕ &h LU=C$LymnEP^)KoG]D1EZG W mXi!4xFN qHcIβ TB L[w؆ ߧ*k?++r =C*7|F)Z^|M=ꤽ- ?y^x tYi/T^i9L45,b$H0JFEZm5AU@[~Y"5gHEիoaҚo)/!S]p+ oGDƪ8ČV`C =C,,:/?\ j{W1$KZmzվO V8?etUN7}0cSM+Ec]wFxOߟnlwI~u}- ;FD˥3zs:=7y~y~ci}f n.;^ӓ߾\^b&x\2,V %?eae;4fl^xBTKˢJP+[]I82z䜤EiV骞%RVaK;9WX_kI?BVǮ V'y=mß:`1ۣm?tkߝnV\;5媞UoN~w6y|Ny8ɟO]{˚o]k`.\~?GX}(zr@_Wݞ ޟp*챩UR@ZVՉ̝']3 W ,m&#}ٿ/7E03KLsa ^~NA 7|ن\aqn 0ۆo T=lF)̶%Q?l_s’&rn^wpP23PiT3 Kė^) a 2/M*̬R4M*gs3JIK,VZ5S`ZCd#Ҋ&rA!fkS2ahKک&ob1"ppPg)0(GC지CФS@i`|`pBɵve+f݃ᱲ\XV\BࣃR9u ~h#;I6t%x Oå֙:Vb YttWB3q ed2Ga ,n=ѲVbd ,Ch1z *Xvm}HeFF(,&nw0`b9JjS( Lꠚ+H4-܁ m0܁/܂ D9R\kwpLVg@*q Ar-$"VV@B(kox o1 l@t^US"TŃ,i]+&7 *{U(O/dwV޶z84l]?JH4$^CjnNC0O--89ۡ!a I,Ē;YcRE1 ѸZ90*ՙ8:2N'8:pKYZ&@F1<k *cYזQX(0.v |eJ@'IK/ LZ&WsL(7!ަWC$yi0<1iIwpP1!Om[K'-eٚRvl>LTNӴ( alMMӊ*bNw<kJm? h V/ݎnaN%"y;,[ƗTiieۏt臽i\2Y1zzKdX;8(֢V ّg2D2ߥc4!!e^ֿ2WSbwf;+p+h9y#"7ѷu^W }u^'[ m; eFła4 jɰr8cgP)t9:/e>lN_~6^9{^ɬ4`ƮUb a_Xu|=m N|solZZ4ֱ-H#,aO~[esL2;^8A.u ɵkd N]Hu+KqQYH} "SKMY-h͖mEw'QgږB/7M2RMrqSr+Eȵ UD3֊7 DE]n`ER?b#펔Rtf2?:[/,ʬU"%}si[uJ Y[(:1YbTNDcJ`޴MD jxzPcv9Xt=C&A|~g>b%}+_mVWD۝*od52fT+TjXWXVh.:bΑ$Wf*aY&B i@=C:Le'NHD@vz5EgC65r;R@q2 >IIK z jZgeN^[Q枉f5wQ `_><S\ٟ~۬ZiUvҁR'q1Tua[RmlM-VD-:+ ,e< T(5i7g`N5=C"8)\vS:K/\M|}5t>j/=&2tlt)eƔ`z7+%봋b N2A{Gd%BTZA3UU L</^a| R~ܛ͉a78W dzBށC; %M[˱} $%BMԁ ,+9i[T~!Ŷ, 2-hq~PKP9ACs(_~3MYr6L0FG6LLG0|ꌆ၀f\$>I2'­>}6[Z^= L؃ :мcU~'cEw|=0O:Kdڍnu l _59js,E2gA$VFGpS8I8!Nk=T$Nbpa$":м('y.Y]7  EwB6P83ٙ^K?N7~Ln7Fk3ch<*&W{!ȧV] d+lOds#a6U߻[ZS#tVܯ9Gsk΋~Mъ!gB(•c/A ces ƃdD&5& M*[M_Ȋm9^+uօ vZEAv{RQTܷDso <RZ-DQLf sl=ɠޱd +. ޥ v+ڮݦ}y>LC[*bG,I-##EnA%Qbm"c,gbLSi {ѕ5Rģ̏84:g(xkPᎽ¯) ja?Ԩ6(3 (g%j$v=fzDcؿGRb@Y:Kp*ӘwZp=!GvchQ+A6e^vk(amCi,. NS` 6.f$KDaoEUϴ GZPqjd0ƾm<ڢ.ؖ2nHXL8#Nu(g<5S)#0=M-5WV@nd#U1YP޲A-fe;b춚2"_/b[n{g)QJ !D[k13leLi(J[Z).lg`]!lHF%Ҭ}uz0NڪM_kb{PՀx)TγgRA4a-8LK ؅Ѱd0\]:X·K,av~E 9dc ߌCgt3Gu2#K>8u_wV㌇O8$70r%h9mdƖ^egEv e&ݯO<B((y.?%fh%%Xtd_(1QiD S|3̡5IjEU8/'p)suEg~O5W:wX8K|xOb!"jQ!xOA<2t Ip.1N Aӝ~t:bjlb";PXQJi.R<43C֏] 9UaI]! 3G SEHuZr&Ҋai4;:kB9@`xa;kv縹O<ee}9~׌?%@L,H"3#dw8=-gd[}~'|_~xt̷F-;TYM곃mN/r%3 is SlelwjUt ]c/T'pѠ} p Y#K\b4&)kcӶx)xbB"Ӝd)bǪZn>[5< G-eTI{J1Z ƻβbl){y}.?YKr>k{r.zVP[fpL,q/ =&X 83F+V`wF%$qWlӊ=,FJ4yw/?-)ao'9uήՃۻ"A7xket%I7(I! ny<*/YjߑJ.4pu=ҍq5}-`t3s,?nPla{`9qO2lghBh$*`_CUxNu(ai!Pi?O7yabOF07bx;R)9pvu$a[h|qlr`U6q0QQf)F(ku^G.e@bb|#Q ~ہU]y`u'Vϧ*!ƨp;"%RFR{σj"$bfTƒ:8<[!!hl z֌GW4!_vek}:Cnpi؉WX ҿq::xq@E:r>?Qw^Pgy_$tWeUt$QEAxG(0( QvQܯwD/ rb%h ώ!fGdw )çs8#g'0fda/&?s3H?=w'd6ϯa$i_qh<|X/|v|6W1|}r矐qVwEzNee̷qKqnevP9"7Wnz_RM]:]̐׳׀'ǰoY\>5Syl )59oV&эO`-͊LNKdB '1o,5z~P"uoo*fo哏Ǔ.4[(_6q>4;B['w+~,9ًW#簟H7_^<&9T7gws;5W _F5K?Ø/ix6>-^u6nǜ~n>#MJOh?tgWqڷ7SȌ/>̾}O ~ SU\B?jߎ^.&/q̶\8,y1UK,|rv_ ٩o܆9/5xbYbuX9EO3ם 4605=XEMҘmy@%z"-yqCGZ.Y<:ҚY߿8 zZR8ǽpg}1k)…w5ո7}\z۬gm O\VR ։䅜ǻ 9Bu4ԫMX䅋(^"c- ab#=FpKݡzjIL'0Y+^|Qgw+"+. FaCt! vUzlGJAY%uH tgjkGȴbaFv_5փz(0<{4$-*~R$kQXCB_{ ?Y`?Q $g;pNAɬbQoXMz:MH[(Byb|y|y|y|y^b$A@"8H%Īƿt<:7Ƒ"bQ/_㎊QՌwL"V%plǽQ ٚGpZrfeh4U=xT2V򌎔 AOk̑tPUań zXĊL``qy##Bh,y$Fqft\zl,=P \@cdK0^Qiz@;)a)eH$^>r` N©[-V*#7ڿJZ J,vP 3;,HZ"% ʆ4(fSƚUʊ1A-l g6c zzAAH@@%f1N3Y Ku8IOr"!ZJWo]>*jRyi}PXP@R%[+ ԛ} nMަ/L8l / /uj ;_IîJ`e-}8;D% ^ѐ`)͡.9_h<-oaW`;pFp?<7]%yRa4voQعЂ8&?̈́ bҖi^=j\3v0S؃D9Z?+d &{o?AE{B`4w>'̒d3VTT9^Cp{Wt_q'4 n&Z)u jCho\ݭS|QtRT1n[ka+-ki(8u-[w8I/Yɼ~hT.tMS ym[$Ww`ER2dq2.#Ȉ/TM^ 6k_v#feg!zCWP×J0ʞ%z.GR>j_mRx ot*&I֢(`pا[vg![pd o=j[[ӝk-Pf4nAx7_ؘ;\[ +-mz7EKB޸z8=2կc+}w3 μ\OKB޸zͼ1TI/xHeMhHG={R{BSMS58b{+XWh2;+6ZST+TA|R xn)<†޿iblzwv]6ݎL(d3]z$3+7{q ̫ "wi|b#<6ewF+=PAJ';rݗ9K'>~Rhp_¡J(a!U4SC<_/z@kYQ0X$*1+yr\!W9Cr0oǵCR }&LԨ@jϵ a\)fTZb w #4°3=}ƨ&VST`&1yϓ׊壞U,78%+EKBV\\^p.$:94`&+2 GeÌp%+L +^ "lu+MMyq LtQ>|-/sȀ)N%H)J\}G0V> ;6oއ[@a\J\|7kJt+ y$Pb[\v|+WyWj g;pj#6.`:2fS5tUZT^P=?zFF7E86 p^VZ2gLhTzXety.#"EKRsHhQ`cIR2AD\Owz܇KP"`bGeK9\ψˌP:DL@8y4AB D9P=_3fŌeGW}N=&dI*3sg",/RT 2K1h1FH"9a)/OxD1E>bTʪ<szHрlvs\Pۧ'c\6B.~飮 b>0Mh!#ޭWKbS_y翿PNel5kh߯.ד@.3hɔ{֛ə`HmDϣ3cXɋb+_gij>$@7#T?LH?Mrgn[@hl|)6 *3E:0A~<2Rނ̱&Hp/a'ӌP4%1SY|ɕ;W@Oڼ6I)ͮ6ݖJQGJ%TW:+՝ZETPW;q p8Wot}` vvw}NY:ot-(rvڸW ,tJmf@2@I{p@ƬBB J y3 91RncqRhX*# Ϗ軉BaQF[{!_iK IBYS" # 32oDL9aXP,OXR B<#⌀zA 6?p(UܬaA* ܈5!~{ #&67߻C,ta򖛐b4h9Ej֛n!J Pnr8L.q<0RYGy"0(Mh$@u@\Da4+%ԥ`t"_GؑP9U!JhH"Iv$(U!FAhJy$H2r@3:キ#31 V_GA#HJB @ |p.f`9ȶBߓ2V<Lj[HP  `3)Rs$G!R@9Q.)D .<0EX1&H pJAvʊSgTM- JO@jߣwQ>dUmS-|v4l^sۼ)du()&һȯdc5HZJGSt9QBt}DġK`eH+aG6=2V W !Zwssڻ9Ւ![ﰹծIҽG*hobA?ۃ3:Ex"H]%= V߽,3/;h  47yAӝk@4ʄPU9A;RdT9M~=>胈$/no6WDXL8&ד-6?iNԿ]Z>ג S]\UOn~|Z;$+"[?Ryz3C[g Ú_eGTR<)=,NTZXdu^k1TYx7n^6%z7 lL.ޭ^ԸNoۻXzMǦ$E?&:n34+}wOm{7KB޸6) q7Y!:?_YE ]MͧsvyܪpN'D HuWB/H0)e_[_@})l`eHjyS|vQKeߢm|}^n]5.iɼ} Atz~@5r^9xes30ƢjSzY74iɉ۵)+:%h9vMO(|nwRm rZf88DJ볾Q 6J9ױ (0-oxQ+N2DBLIޭͩ18BO%+$w1c&;+. i"' "1.D 鼛6B퉼mi !IqwuͽuftZ'#HݦF1bݿ/Us9 th^N;kNQ R9f~@53R+N|e`sUS]=eI y&eS>}?nINW6 :n0c-q)N1x7S>=e`Uj@ey7fjMMa( U6p%tgPюy .MOS=R-%cYW"fgi+ JOQ=R-1c`gh+5STϫTK6Z[) (.KKRIY1ZY)V-`9oKOS=?0"`g1çO+EJ %AbҳRL@aX)vVZPM!agDٲqlاz^HG_zV-sE\`P-;+,Q ;+-Fhhv8G1+;R bagXǥK9+J %ctVJY=aĥT3KjC4P>,FGzPS30Σ1A,3I$JrNc(#H BV@ATK=KTTk"& oOgg*OMl9/jӭv^`POPz?q|\_OQ~fY^]^Bg@^ .e9VW@d,'OYyVQǵ.0.r?x~2#N`難6'[@L%lx)E) s2 Y<|K-r4xWw2P? gR&hwڰ1 >z>Aq~TĽE^vZO۰GA -׻`{R禮ݞoimg6@D-,qr2*Э!cMjOPg,!^A 6uv5G8V' H(%22DQʘ8dDOSTjOTZ;r 5v2cRVd-ug!jT>׶.={,6Fފno;MR!3٭(7xl+% xu6DEٟla|$/ 5_B(ɻ?)ӉۯZO,Ԩĸn FU^lNJ껩.heqC llon) `0`#9JLG_tK@>|ΉR!ʕzdBk!Vw}Zܩ?aSִ2eDT6exJ"I@>e4 r3m>@Ғr$) >mTt(\EMƎ(fy* 3f '}![iB(a51nUJX VZd~$蕬Nx`E C k`B )c.& @ckmO,SccMDIģ(8#y1g{FWb&k}2 ra<a'仿j)pSD* fW(~UUI8RJ_ |דDZѶ<, XVPXZOaV(@1p-GqWH10A="q(y7*ʟx|![;RS%Ou8+W8ZGpr$BTsr8b. ֠<&V-<*7 ɫeHZ2Pv\`Pz\i$$B #Y2X |ΥT-C_=~lYha37_>R\vg1TgQeyH:Y` |kEYdhލ՚sQE>7>1:9p87ŝ\+<-v'0q{,כww=c6nEluqW[x5!V}`m*x)¶8TL򑯘4ͥo%,gpp*ñ(q|Ă_ώ{o<$VdZ9$F2,,I5%1LŁH`+ X?޸ҝ]YpnoRr,+Տ͜&Ŗ-#pL>Vy8mTp|Lw^ σ2J8kL|Nw03Nz3[F|vފ["h)z ,[|QL.D\)QY8bs΅*]Voˣ.dR6({~0}QR#(3] hąq/@&MSulYTkRSaYgVZgYʲ%~=;aٟjҵ~*UT6җTTw>oLhjY+n[IDt_lsD'c|0P5`d'Z*M|bgG@is9U`2޾(IB)"d=,n&/gJnlRD1ŝ3݋fVx;+VrCD>gR[N-mCNoQnt^+YڑrHv I! IU=K6JWA[jC[,\_AYd!N9yB_\cۯ71!ivIە$ܕkoyGS)6-Pl1YXݓs6qgr@' G1]2qO^[YQr"h݊= ZץxEAh/ ~zqmC0 5!A.+R"'-ld*"*m"wvB /a#NI9  akOZcd-Q+=ZѨ(3`!F" j)rcp1 8=B\\Jq$\y&a:{: ځ }`M[żms3|a*[X7G^|: V"Q?pA0ѐb}X?J%买PP\sŎ' 8ŧ03jӲz f,v?%=xRH(8u]%U<VZ:p m떒zXKf7R)kYR/K[:k%)I0ٳV%b zx`**J-Z6^i|/fP%$n>ZѶ\h vx&E⻏f2J*ږQ$F5KvɁ KYQ}|G1TzAeDc~%',cvL@'uzmXSɬu9s'g){  Mga;9Y:Ea#DWA.8En.&,@JyL]hTsXx_/KGw6SNLS[n6Sl6SmwB"\?Dԥtp!s]p+Bˉ}%d^`,C&c|ZT^V9b^30ε.;@oM± v=U,4c8cTޏ&7$^)2{=cYu5YȻ/,nכso]Myx7lջ֟ot!weK^nd#;b#^U |-F;/_$ ,BC~[rANR@{eي,({9R"tx- ) ]"Wy$X v.H)J)vRJRʽRJwQRY^m;)HbRJFTkq;)efȜ󭅔FDJK)VP&si>Ͼzi}ޯ&7&%Vq|fdJ2ۯk˦SAJPa$i]$:DWICMUFs.Xzz$t)_^IF05m6{mh_ =[YkbG@"MI;X̩"Qu$Vmy"\=VhV<4iU@#5jUNlӬ. x\#']YAuFXꪝt{z ?j#V#MCas(Bx>ϙC,Bq'rBz!K-47 ;LA퐭fZQ&׻g9Z3|G  YT@>%E!g*!r`2acoI)cis}ZTlt ;x8!RNb~y9}{% g7?"(/&^| 7aoJOc)i# irwW5biOx|y;&p}6H Ջ/A"{y}ۯỏMn@.^'7 dyoxWp& iwWMWWG]8[|<"ܐ{sݒ"=|`GRRd4v>G6fw,}Oނ;s-+dtaqz6ZT&`NŪ%h3xE8_;ΥFW.Zu<\z7zLnM!x?=|6<&QoF|p| =0ed.|rƎJ鿾?a2<ի07hρ?}=6tLQ͇ϧg!AMF÷Ѭ,Bc!s%IA-+BӤ"t`uZVNք4SrBPDpHśpnM` 7&0\]`8-5M`ցajb? i#rh6 䚢K'hFD6ai *4G ; Q8mag5`#;#քZh $PM>hҦ3aCi EV*澨(5e7$NYcos* q0XZcbXTU*5jM8Ɖk ,VQc8ư;Ȓs9Fc/qR;ipq0,(^Ec0K$NOwl4)j 6,jx͐H5OFw}?ɩBcmD]Q~rW̃ ?Йf/իhho쭒I1IWkˋih7yxM^5yxM^yx9nZmó-99%fF&6q.}oM˳Ow̌"hǺ"L7jھ/M۬;-P ka8 x@ڊ{>v!;)֬ Gߒ |X5s[02AF0,;iWrI.` 7>;ʁMxk+VjT˹w|_5eI ěmSIJsv 'N7<|cM>L,v@vM4I SΟF 0K=L#+00&J{↾T+OTYg\;1oj8ggkUEp>y"6s͊ `!ƭdPi?okG0金<1Mf@W8Ў~t|՗o;!uLi$jc" 3~s.6oE [b%|=C?S7$ȣE&:%AnܕZK6W+m@k(.9 g8"!+[6^5$>FLv:1$`Cfg0aKW4%pٞRMb{ƣPK`1/g1$M f#"e,FQu;$V q8/V['/=_|;p#Yלk"XqeB|OwyC[%"(8cUa+ a#Pq0DaH"@-bG>yKJՙ {tʝV͇u O7]^W"6 $?}B cu9U0{awZ5M˅}7UQζD(츣J~78ng_oh[VANtCX=CbO('؄"4vlĐb4!kGgøT%K !zH v÷`K[֊@B\Y$tԘ2a2qci $cՄŅ4MkUlk]M|rOgWmr|'#Z!oQXzQ|uYkHz#k$$b$:^ߎ9 8 [݉[ͫT0H#5C-lmx.k؉M>c3>c|%U7'3;mN5M>BbRܚ-P+>|;^Fe~H XPh,@CB%SurDŀ 53kuI7ީE'χf֦9 q4Of7y\0%;pbgݢ@+LF r8DԗS+S5UzzE&J-z& :dܯ;Fо:`xɫ;#ڄcLP!X?rdۙd4u_fbӀr[[h=vtE]J xvn.| hbc2+-jA6ѺBiEfCЗ*->{ kCveqD@˭=ރ,d32B<,&nDaJ/Ar㕈BDƑ?˂$|l:57+JIiҊֵh6M畷Dy^Ց#QQJ${gx NwܗwW!+Ym$-͑Dm$h(7EbMSe)֝Ky]3/z9 B۹A(F-u!Jde `);:`E B(<աDMCr$@>C%i m~V >,ICz`L|1 jHD9u\TKYPRQ&8^igsAzS;VKCa}eI/i0G%$J0j9G;3'.Lbz%>ȻfPdҖVR(E(N+.xǍ5ge~}Y矶cS+j`:'jni+bhFDNe]M59}[m!*fuxIAei'΢xonoI7Ds1'nUi:UQF`јy4V=RQ!OE+qޒn87ҭ* \X;HZBXqҭGV=JlVq9gjf~o Ag4]Eؗ3ՙ,nv?/9Dp iس0uѴ ?z3,F [YN⛋ D ݐ7=gsU('(Us6طj-M@bj' _kAT] u,Ă8*ebrX4 F_P-R.Y2SQ)Sb[>rig3A9, Yg(AM%2+AY,=h˭U[ .aƉz':GuioƇ 2%ra{j>=y^R\T~:cٚO  W`"|ʇh ,pG3ȱpqas*f.:z026Գ<Z ^c+";|Ynɝe4 No*i9|Q~Wx.*MYCr2 &OtzCӭ{č0զ9cLSXP Gkg3={m =kNͮ'g߯Ǒ>_Z4'wv)UhїWOc+3.NFo|U9J,ˬYf0eHñ5vn~t}]ښA壵F:f86v[=gUaP1"gR>^wC  J4IT@e/ H!5Ro{{^׽~B{Fd6^ofig3Ao4eN[APh.dۈu؈%@PۈF6mDo#F, ۈuj235g#eفY)ZX8B!d>Ts \ULA'/(,ٖ'ʯ3Q1udQS .hJPPH  4NXv}.߶^mYmmm~ےD߶~[ qVդ.NK bqP\rtP%2C;rPi; hK3@)=#<+7RQz*ˢe#$Wx7Wx`* @F{W}-r:¼ ppoeW:YXJuCwX&yE\kmE|;GoɑR*"gp9Rf0L( L?0ޛ @\m]n NYK29\ӕcY㞱 =x@qtfr+FCޞQOK[쮮P^b-qTq4F3.A䥐#KI m8vBcHK( 5Z481.‡T LsX0-kq͞Hr! #_l/ى' dkDŽ3l.Ϲ-(\.tEBӾj7 !4=BCht%mAކ+Rh!Y+=K5E_-g("7d~m Ng]ֿ`r6620'}J䪆y\GP{+ 2^u`t Qb`ݿ%wXB2XX*8Cǜ%D4!,d)6ъH.&>[}$>Z`KWPG}["Mp=S˦N- )ƄH/D6]̡Zc¥CI~D‘&ɦpTSwKlP P@1=s@SOC}jP'B9K)w̟fŠ&*H5.+%)$ xH 3p#(5\kL;w7FrŠ@ q]p6 hΈ|Ndbl\&4 c SfB#y H0*ĚQwpht;vO??GdžS'|sz`_w{?-0u2f7b6_gB:虱 W$鍳:nB#L]zQ+Ww2>+j~E)N^j_"=yW3+U/F`yJYz^w-I<g=,Ӓ)2H!. +>T+,ۿ{f,B +@,)(KrCz(ɫ'_#(nBZ]dx&Yۓѣj^eoR+&&t> {ъ-]: ꢥ982w6ۘ!XD ) >zQD`f F-}8Ro;"jpܥ 44sKB*7D֜oURjm;*yb(H :c -[w&1l7+- 礫?16+Sl@A#~hLԪЅd%liP-w$xwM<jԽg=u!U}]ſ*M/l]ú6}ţSxB8p"vF Fv9nA6Yf@S/#;7}qgȘ s]I^[\SyXJE«z8gQ2Rj=fg}n8;HW[nIHV|,S'`qI7Ktu|QEuEXFMkͿH[b9OQtZ) M \I8ɇ#H9Py')4lx BCcӊ|o ({&wl}.lFaLܺ-C32ǻLKUiw? `meF"W)fȕ>\i,⮁Qni8|5Kz.>g;qut;^/ϲ6ixzۥ(9yzrjlVDגJ x*D$&m-U  81i13C'lpL0*dj(hxkIV uODnFPV$ ?L1II04d,5LRcj|DČ# Yr#H C܌ Ҩ?"J?)ʨ඼H@%$LZJ\;a2.v*,Xk`+@%g#+5sVaLbJ+[P* lu#EC"/"fTiRE Mg͚WW%fZǭv.zvH6ctሢKK y#o\vp:"nm* >;!JhzQoqemޖSemX:ev,Ԙpa_vg`]8ݾw^ #f>~ZP>\JЁBsIBGQhaaDžO)!:Qr8_c WD5;Y a飐&(IbB3iMaBX0ǤԑPPdžQ2&ih=QMy#OێЉmݭkr0jJlBQe=C<ӖڣbCiKUg9Z)ϓѣTW4$ IB8b؉p &dq4D8/jiy!(\+]-4.wTKΠ2^խ͌KCq)y5ɾtTFwT:6n}U򍳨OsOqn{!|qjGJS:"8|^߅;!8zbuWtQs⥊zVD'zǷ!+i8E+-p{\dpC)S<>Ui-Y*:H$s5}|> ,12nJAb4&1I  A52iK &qfr .$F7JC@-뽐8(HCd6dCWs .ct?$.K2Ӗ-ܪA! NbH[+Aj+PvCO u9 :H$b$/J|=lӖCU2n$Fc)`Wb{QW+"CiK kkVOumg̲_F09#WwMF~? fIS"5&A( Bf b#H, NRAb (XL(S%ʦ<>8* xBXn @CLj̈y8ܛU!: a: J "bZ4FpbۚU&V(峪-4(aYՂNӡQw+e3Z(1VjжfUgb@IΔqcc4 UӒqty`NaA a왂([Y~zl1[0yg-vnb ݌$W`+BiD^jAC6r πIekL K5D3G2krH+JqܐdodLDwqh~cTg~?=P`~jVE?ˆ !J \āV4 tZ O7 a,Ts+JaR1`iC)GƿCDC(8PAX(D*csR4Ji  &HydPkə=\Amu/f3TUt"/:9Yԋt5ɞtSաtu|QEuhMe{ )oExJܗnءtu|QEuE(-ڛPtCq=>*-0TÙίOj7jY|D9yw #_G?.zDQ m{grm',pl"_l~¡1'ýDL2 gQGA*uDI+d uIrN eUxFxŲk4?/[#d?dU"l\ׇ{,LQ5\^i2z[fa,@D'm2*]k$E8d,_ˣP@iYׯL;at)kY3;l89=ψ:9vcbˍ _~/->-'Xqcyrx $^_\?/}swy__?m.~.&oo/ |=W|䜹s5]$k׀K ]5 7yne&3m0 / 7ٗI̖yvK.e6d>?pr[ye [ 8h<#T-\2le-%f(:z.o]uKi}f_?%<{ӏpwmI_!a6R~ Qp==.#$e_ % _Rg䆁D"G=]Ww>%sxv0ɛ|b)ЇG N//x2wHxg#|^~aӊ'ߠ M iWWWoJ4n9>ǜ~E9ƋݫΦ.WrLnǑ'\>MsBq9o_\_\\M"2ח?=.>"~z[2j*8pWW/ez.Չ{~#̦z.1/CՊr\I_9KYVLaŔeOaVLYnԈtmag qr:bQFveo)˞d,{29{5m@A.ec\E"2&&l,9̚,(*酈\DX*Ez!Oydb܈(!:j2|DŽw0G#ו;2F Cn\kҜ!{<܂)W)TQdp|eue\siE AUܣ !Z) !oK**3~)`ւu=ڋ0P vac]ª/cE@! <.~0C$Et4]wW=, rp - ;|\#uz89lWNv|GӢ%lz-`$(]Q_ߴuԵTlznŜveX\2k2A##4"ʥFĊxYX ]i Np4-… iEs dw>iNVx"yF$Ր=jU,|pk+$D=yRaֺ0?^^a7WI"5`ig = P;P^of~RwP3pEU7῍yy׏o7/6P >9!4W->s_lՓ]c&k7=YKF?sO!8ѳ%DS"ue«vQJVm>o5r_/h]].C/veUq{80(ކث+s@f|8!0k$hC3ō.>LwGgUa~t)ؚIRk B)&z(s$wJ٫`^$@Bg|]D _Lgfߎ럌?&S=[aSu7@/>z4?3%&`i0;ӕQ_a%_- j]zTe8{@4hbqWu\5RR}HZ`lOgL/+>:MMf&l[w@КtVn/^%rr|qyG =,]\ N 3CP*+Sa6P1㯭fY%Z>10?!Ovu6攈Z)1x__dO_}vqs=\R?8@oinʇ4v‡9d)8=*cF8h{ٶ5MS,';&hy5cҊfh]OEi҆$ТRJhy*PD#1] aL-*-3ր^X]xFmnRȔb bFӢ0h7232J2.cu4V֯uT\Ã#&pR!oәya:~jp:P^“8 7(304~3j==!|f'ǔD+x˿Nƃ,>Qh_rnc!^.!Wچ<1zTvlſ)ہNوQ#e3vG=Tˉ/zw~.X>=b!z^}>xuA@)qǃӿs!9(ml:~~fpBh0▇#U'w5g78T kp%+Ù @p¾ojh- :DBXd !U[d g H,8/[N.?"-҅}D/P/B#61Yz#v#[| $Rr}{cup% Tn/[<.@r&T\hȴ&Nkȭ3CW0 8Ay24v5X@0Fg,D xhy[n.DЌF'pi[Sp[,@((9pC w,X8A -؈.?(5m;ZR5c-BȒ!9gQq"ԉ-߈zDr) K"b-|0.'-H0n8ձ&E)55PJX\BJ8ҽP#%@ǥtz4[沲>o)-hTnE䶗=&\ Rn( sW!js,G|*~[Β76W%а= IC:,K؆*riDW_>ddPV0DCmL14ɬc@V -ZNkl![2jKKۢ`Z9 ၃nCSB5sݖ&D:(͐1@>mMN(4F A!gǯky<  H!ѷG# J biybZWz*σr(Aت<~Y_A)A@h:>):7Ǝ/\r> r9MίZeު+ύւwE$HḷEFQ2ͤr^:K 5.#|Sj|B,_ɭ|ᾋA V}t=I8˷uVs0^ZX/ZzlŴNJiŴN.-xM5Iւ]6cN܂q\zE7;K2igm$kSE -X[3s+)e["JLCf pDy& AbtKF8I),-+Jb4GvRtӗl#EPom~s ♈B%݂SyF3+mF.HȪZXGh.<\YQ\?NZ5CTzMaҎs;Gqy]iO/hV;1j!۵+CT}vj(g:qޘ%kelF1ඁn7!CA2VPA BgɈeAh` ~Y+oڊ躒o?k%iBFn}=}GPDwM/bgؘjPjFbp9z40)[UZʼ̱2)enYkm$WŘ歊= H4xeYxS/#{d[-kƳv|ZUbWUQޅè͚]1)!U{KLSiU,pWEѿ_-ꌚ"$7L߹?zˠ{q/|oI}?{*/V"NXLX:mlWfϓc:F˺ky{W4|peSjݴX.ޭ ژNޏ>N0{nȵnCX7^6 ʓw+63x\Eh#޼[omMM6j[V'mYok ۲vovݦE isKCz}w% ]mYܩ&:Ois_/%3iJf%'$:/W7&il!RI_s %NAơ%d̵k̗YFU`8U{|Ҍ]~Hj⽥2vіʸɷNe޵y2vШኋC=SőR38x^CS.©7{yvg -R3fq48/\kYCc9P{Fo1Ԫ'7W};s9w?}}KJu?.WQkѩpo`;qνMZi9QH)-%b2UGSCܳW>Jѿ#c;_ruΖ!1/gO4u62Ύbu;q9rE%+Fͮ W/VxHxkݒQXFmSQhMqq :H~4 %Y.o|`3/N{OPM](Sɓ6ڬ+$ÅX;Kۚo~~@tU~i=>?D| jFF'yJnmI+b8"I|R@IRיŒןo/OA6S3OG'([RάǁxE@zdz" wօnhl`/),!19=%DCqGx״3X%">7R-W ;D7pvwO jWu?v??mu 0lwqB !9bhJچT=eEFȐr&dցL+2s!2dZtO^ ܍5V^ݼ6jtՎO]ë{źȵRd"5NU<ɡt\M j2O@pc18=k"CVە$W2VE\{>5AӚmN?te@\,꜂pReqCyԚԿV)16fQ\i%x kVw@RP|fGۘ]\Vz$aӻ;;kT$V5f} NeWf 7ņfgCъgFNT1`e;v,&;ES4Nj7F 84)fxi2*9@%c!ޭ3/g_iqSY N"C^a%6o+I~@"챾1iDTs#9Q?ALr"-۵7ydmcTHMvA&-b/D2E*ђC|kyHf7k[,4R$\B!P:E [C>&XeA[Gb$psPm޷V6,I+Jk--Z bn b޿zZ< bvШ= bOcl?+q=d5% |ܕ*{CI+a(h)KA_`<4£Fn+I[ЄyKQ2D*K?lytP|GU){ }n"E;<,l3W_@^Rős&p)&tXAk0`x Jq!SٸZuF8}HTM-MxQµF&C"ANLSAK٫(zތ۱%zc%OhSKr)Kp:u8FSxkY)!h(Ǩ;+x wf6c?^VQp ZJҼ_| uρږB>)!GVxޭ ژNޑ}N ֻ a!DZzwsB{ޭ ژNqH؛w wB>)Ҧ1 |DmE"e?;!K&$,Z1P4oݷ&XTb4#GŘؓ1WgVXi' hy-%aŤ*!h4nu3!B5+a@2_*S u X!}V1ϗ 4$tE[)\iD&鎺0]۷=o-/`@)2ܪi5.4|kU՜yHiFf%L+">ya}yFaDž- ̜Xk*dŔSO^xF29P|zYLn湸dr5z?]7ML䯸RLIZ/Qdu/~ x{t#|M1inBX"]d*z3:fސblD9y־(GFYoc3&Ձʄb8F%fdq݈֨ecUvxJ_P[P<F.Y64b4L&,.OX kDH3H>ݤ&I)qVIȕ`.kY/'AKB'/,[qc yUi2_6kVɖ)0u4)P,90`0eQPFCJh%F]cم]0 pl4aI@*$b&۔FC$X}WE_\Wu  g`PʧF(ār_oni+va,\_pj!~=O[ BzFhZc@U2Zj?VjSz%ZlAKxp"eɡ`&Y_80cǽ"$̤b1ꂑ]FdCI9qo!&fH>]qb fJz |xwhGA&BJ3+`!yMxZɗn#8VTm&T)6Vo~că*QlH!8BgE?PGVuMx>WA/jIuH@3LswSOU,;fطu%tj%kFJ$A͟:FoI휳LիYlh_4q:|Qa b8QjaČBB 4BҊ CRc&U"_eqFZg;q8rUc9>@/<>$YēQi׻tG V!|T J)bkr~=x$Arg\ئ&84We,oLF;N=C[jVr}ـhNY?w&a+"膐 0)TjUD#Dz>?'&_j29jNSZclrE]Kg˝ڼѹ= s =Ob!,H1)YY"b&|qTP$W )hi$eh%*`mͭ Yԛ[+ܺqFFo&34&x0f\WP $EϧxXcDR'8}". zl!sra1`  ,u☒ dF(JL]DgNx?FFi/Ʒݶ /liJ* yBX.bpb[\Av"j 8Lk÷L?2p[mw(vo~cƟ;h8;Ke3+Ո'zVY~uу.v|Kx;c)2)V%z_sU(̫t@{vs:8K||I8dؠFt#(]><^LD!` Ւ.-lwOlJ9`;z΂#ʱŒFj+o3LFq,bHlBͱ C?,iB$E5Rb:siQ< <$EF&zzv8VMRq@:52^m='mڤ`aQ4#g'l.$蘉Kkf֛r^a9I[FtQ3 ;}ZV/Fu˚d(0:ߨt[[xU4=ɟReaV7d w|J=[cEq4|Ț,UB*y+؜e~K/S8[ky_GvWᗇ'a/oԴB?>1o-L4m=^ar?Ǵ?aXx Z蹜6yxڽV?Mլd`Ᏻ濝қR{'i{?<B5{AE~۾o~{hklڼO m:tQM_O &ɭ=^@\Fbt\g;fO pC74?w L{0>>]zqbvۺ{gnj'Z府K ~^c#s+ߞY,{五"ORQg]:e[c3Ul=F%FlF 4rOCE_7Qq4s dX`̡b92> l>/..sNGs78obA⏕dO1CuQuQuQuc,Dzt[ ಉFS5\ }TK[HhV=ҧ>]zz/QK1r&"xUT>!n( s9({Lr깜z.˩r깜z.gs9+؄IR\`DݢdBmryw%ꨁ)Z:J~Q2nZJSZէVZT:-\RMhL ^^؊>]Z\VD[}hNVP;AQۊRPSvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004175550215136702720017711 0ustar rootrootJan 29 15:00:42 crc systemd[1]: Starting Kubernetes Kubelet... Jan 29 15:00:42 crc restorecon[4743]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 29 15:00:42 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:43 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 29 15:00:44 crc restorecon[4743]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 29 15:00:44 crc kubenswrapper[4767]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 15:00:44 crc kubenswrapper[4767]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 29 15:00:44 crc kubenswrapper[4767]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 15:00:44 crc kubenswrapper[4767]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 15:00:44 crc kubenswrapper[4767]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 29 15:00:44 crc kubenswrapper[4767]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.648778 4767 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661012 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661080 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661091 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661100 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661110 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661119 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661128 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661136 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661145 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661155 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661163 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661175 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661183 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661200 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661208 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661217 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661225 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661233 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661242 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661254 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661263 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661274 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661283 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661291 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661299 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661318 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661332 4767 feature_gate.go:330] unrecognized feature gate: Example Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661340 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661349 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661358 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661369 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661378 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661389 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661401 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661411 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661427 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661483 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661498 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661516 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661525 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661534 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661545 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661553 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661562 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661570 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661578 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661594 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661602 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661612 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661626 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661635 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661733 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661745 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661755 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661763 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661773 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661782 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661791 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661800 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661808 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661816 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661824 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.661835 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662325 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662340 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662353 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662363 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662371 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662379 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662387 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.662394 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663655 4767 flags.go:64] FLAG: --address="0.0.0.0" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663689 4767 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663715 4767 flags.go:64] FLAG: --anonymous-auth="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663731 4767 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663747 4767 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663760 4767 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663777 4767 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663792 4767 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663804 4767 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663816 4767 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663829 4767 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663842 4767 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663854 4767 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663866 4767 flags.go:64] FLAG: --cgroup-root="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663877 4767 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663890 4767 flags.go:64] FLAG: --client-ca-file="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663945 4767 flags.go:64] FLAG: --cloud-config="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663957 4767 flags.go:64] FLAG: --cloud-provider="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663969 4767 flags.go:64] FLAG: --cluster-dns="[]" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.663990 4767 flags.go:64] FLAG: --cluster-domain="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664003 4767 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664016 4767 flags.go:64] FLAG: --config-dir="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664027 4767 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664040 4767 flags.go:64] FLAG: --container-log-max-files="5" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664059 4767 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664072 4767 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664084 4767 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664097 4767 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664108 4767 flags.go:64] FLAG: --contention-profiling="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664120 4767 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664132 4767 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664145 4767 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664157 4767 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664173 4767 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664185 4767 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664197 4767 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664208 4767 flags.go:64] FLAG: --enable-load-reader="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664220 4767 flags.go:64] FLAG: --enable-server="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664232 4767 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664247 4767 flags.go:64] FLAG: --event-burst="100" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664259 4767 flags.go:64] FLAG: --event-qps="50" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664270 4767 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664283 4767 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664294 4767 flags.go:64] FLAG: --eviction-hard="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664308 4767 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664320 4767 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664331 4767 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664343 4767 flags.go:64] FLAG: --eviction-soft="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664355 4767 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664366 4767 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664378 4767 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664390 4767 flags.go:64] FLAG: --experimental-mounter-path="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664401 4767 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664413 4767 flags.go:64] FLAG: --fail-swap-on="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664424 4767 flags.go:64] FLAG: --feature-gates="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664439 4767 flags.go:64] FLAG: --file-check-frequency="20s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664452 4767 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664466 4767 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664478 4767 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664490 4767 flags.go:64] FLAG: --healthz-port="10248" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664503 4767 flags.go:64] FLAG: --help="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664515 4767 flags.go:64] FLAG: --hostname-override="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664527 4767 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664539 4767 flags.go:64] FLAG: --http-check-frequency="20s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664551 4767 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664562 4767 flags.go:64] FLAG: --image-credential-provider-config="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664574 4767 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664586 4767 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664597 4767 flags.go:64] FLAG: --image-service-endpoint="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664610 4767 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664622 4767 flags.go:64] FLAG: --kube-api-burst="100" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664635 4767 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664647 4767 flags.go:64] FLAG: --kube-api-qps="50" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664660 4767 flags.go:64] FLAG: --kube-reserved="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664671 4767 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664682 4767 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664694 4767 flags.go:64] FLAG: --kubelet-cgroups="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664706 4767 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664718 4767 flags.go:64] FLAG: --lock-file="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664730 4767 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664742 4767 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664754 4767 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664774 4767 flags.go:64] FLAG: --log-json-split-stream="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664789 4767 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664802 4767 flags.go:64] FLAG: --log-text-split-stream="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664813 4767 flags.go:64] FLAG: --logging-format="text" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664824 4767 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664837 4767 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664849 4767 flags.go:64] FLAG: --manifest-url="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664861 4767 flags.go:64] FLAG: --manifest-url-header="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664877 4767 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664889 4767 flags.go:64] FLAG: --max-open-files="1000000" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664937 4767 flags.go:64] FLAG: --max-pods="110" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664949 4767 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664961 4767 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664973 4767 flags.go:64] FLAG: --memory-manager-policy="None" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664984 4767 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.664996 4767 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665008 4767 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665020 4767 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665055 4767 flags.go:64] FLAG: --node-status-max-images="50" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665067 4767 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665079 4767 flags.go:64] FLAG: --oom-score-adj="-999" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665091 4767 flags.go:64] FLAG: --pod-cidr="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665103 4767 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665122 4767 flags.go:64] FLAG: --pod-manifest-path="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665134 4767 flags.go:64] FLAG: --pod-max-pids="-1" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665145 4767 flags.go:64] FLAG: --pods-per-core="0" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665156 4767 flags.go:64] FLAG: --port="10250" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665168 4767 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665179 4767 flags.go:64] FLAG: --provider-id="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665191 4767 flags.go:64] FLAG: --qos-reserved="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665203 4767 flags.go:64] FLAG: --read-only-port="10255" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665215 4767 flags.go:64] FLAG: --register-node="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665228 4767 flags.go:64] FLAG: --register-schedulable="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665240 4767 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665261 4767 flags.go:64] FLAG: --registry-burst="10" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665273 4767 flags.go:64] FLAG: --registry-qps="5" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665285 4767 flags.go:64] FLAG: --reserved-cpus="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665299 4767 flags.go:64] FLAG: --reserved-memory="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665314 4767 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665326 4767 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665339 4767 flags.go:64] FLAG: --rotate-certificates="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665351 4767 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665363 4767 flags.go:64] FLAG: --runonce="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665375 4767 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665387 4767 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665399 4767 flags.go:64] FLAG: --seccomp-default="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665410 4767 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665422 4767 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665435 4767 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665447 4767 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665458 4767 flags.go:64] FLAG: --storage-driver-password="root" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665470 4767 flags.go:64] FLAG: --storage-driver-secure="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665482 4767 flags.go:64] FLAG: --storage-driver-table="stats" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665494 4767 flags.go:64] FLAG: --storage-driver-user="root" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665506 4767 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665518 4767 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665529 4767 flags.go:64] FLAG: --system-cgroups="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665541 4767 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665563 4767 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665575 4767 flags.go:64] FLAG: --tls-cert-file="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665586 4767 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665601 4767 flags.go:64] FLAG: --tls-min-version="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665620 4767 flags.go:64] FLAG: --tls-private-key-file="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665631 4767 flags.go:64] FLAG: --topology-manager-policy="none" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665643 4767 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665655 4767 flags.go:64] FLAG: --topology-manager-scope="container" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665666 4767 flags.go:64] FLAG: --v="2" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665683 4767 flags.go:64] FLAG: --version="false" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665697 4767 flags.go:64] FLAG: --vmodule="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665711 4767 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.665724 4767 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666064 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666082 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666099 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666110 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666121 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666131 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666142 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666156 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666170 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666182 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666193 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666204 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666214 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666225 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666237 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666247 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666258 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666425 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666439 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666450 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666461 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666475 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666488 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666506 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666519 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666531 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666577 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666589 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666600 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666611 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666622 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666633 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666647 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666658 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666668 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666678 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666688 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666697 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666709 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666720 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666730 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666740 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666751 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666760 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666771 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666783 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666794 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666805 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666815 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666825 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666835 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666845 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666855 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666866 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666876 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666890 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666951 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666964 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666974 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666985 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.666996 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667007 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667017 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667027 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667045 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667059 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667073 4767 feature_gate.go:330] unrecognized feature gate: Example Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667084 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667097 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667108 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.667119 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.667152 4767 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.679800 4767 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.679846 4767 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680000 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680015 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680025 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680034 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680043 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680051 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680060 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680068 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680078 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680089 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680097 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680105 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680114 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680122 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680130 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680138 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680146 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680154 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680163 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680172 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680179 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680187 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680195 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680204 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680212 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680222 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680233 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680241 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680249 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680260 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680268 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680276 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680284 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680292 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680301 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680308 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680317 4767 feature_gate.go:330] unrecognized feature gate: Example Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680324 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680332 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680340 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680348 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680358 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680365 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680373 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680380 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680389 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680397 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680404 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680411 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680419 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680426 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680434 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680444 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680453 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680461 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680470 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680478 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680487 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680494 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680502 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680510 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680518 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680526 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680534 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680541 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680553 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680563 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680573 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680583 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680591 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680599 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.680613 4767 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680843 4767 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680855 4767 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680863 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680874 4767 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680884 4767 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680915 4767 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680925 4767 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680933 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680943 4767 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680953 4767 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680963 4767 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680972 4767 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680981 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680989 4767 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.680998 4767 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681005 4767 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681012 4767 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681020 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681028 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681036 4767 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681044 4767 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681051 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681059 4767 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681066 4767 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681074 4767 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681081 4767 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681089 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681096 4767 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681104 4767 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681113 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681120 4767 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681129 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681139 4767 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681151 4767 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681160 4767 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681169 4767 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681176 4767 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681184 4767 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681192 4767 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681200 4767 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681208 4767 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681215 4767 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681222 4767 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681230 4767 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681237 4767 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681245 4767 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681253 4767 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681260 4767 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681270 4767 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681280 4767 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681290 4767 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681299 4767 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681307 4767 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681315 4767 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681323 4767 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681331 4767 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681339 4767 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681347 4767 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681355 4767 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681362 4767 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681370 4767 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681378 4767 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681385 4767 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681393 4767 feature_gate.go:330] unrecognized feature gate: Example Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681400 4767 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681409 4767 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681417 4767 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681425 4767 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681432 4767 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681440 4767 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.681448 4767 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.681461 4767 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.682694 4767 server.go:940] "Client rotation is on, will bootstrap in background" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.689064 4767 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.689236 4767 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.691827 4767 server.go:997] "Starting client certificate rotation" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.691875 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.693782 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-01 16:54:00.285322982 +0000 UTC Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.693873 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.721106 4767 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.723540 4767 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 15:00:44 crc kubenswrapper[4767]: E0129 15:00:44.725447 4767 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.747988 4767 log.go:25] "Validated CRI v1 runtime API" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.856929 4767 log.go:25] "Validated CRI v1 image API" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.860180 4767 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.868591 4767 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-29-14-54-18-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.868659 4767 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.892488 4767 manager.go:217] Machine: {Timestamp:2026-01-29 15:00:44.888117039 +0000 UTC m=+0.698434128 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:1099683a-5760-459d-826c-e4f968e7100f BootID:d5052518-034d-49f7-b53f-ad999adcfc00 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8c:92:74 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8c:92:74 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:81:c7:73 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:b9:e5:45 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:b8:ae:b4 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:af:db:fa Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:45:dd:4a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:d2:4b:7e:71:7d:96 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:6a:fc:a6:01:6e:a0 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.892780 4767 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.892965 4767 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.894667 4767 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.895103 4767 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.895190 4767 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.898658 4767 topology_manager.go:138] "Creating topology manager with none policy" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.898717 4767 container_manager_linux.go:303] "Creating device plugin manager" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.901037 4767 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.901082 4767 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.901294 4767 state_mem.go:36] "Initialized new in-memory state store" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.901388 4767 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.908781 4767 kubelet.go:418] "Attempting to sync node with API server" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.908808 4767 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.908825 4767 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.908841 4767 kubelet.go:324] "Adding apiserver pod source" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.908980 4767 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.913130 4767 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.916277 4767 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.919543 4767 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921073 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921106 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921116 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921125 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921139 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921148 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921157 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921170 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921190 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921200 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921211 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.921221 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.922757 4767 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.923306 4767 server.go:1280] "Started kubelet" Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.923731 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.923802 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:44 crc kubenswrapper[4767]: E0129 15:00:44.924206 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:44 crc kubenswrapper[4767]: E0129 15:00:44.924240 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:44 crc systemd[1]: Started Kubernetes Kubelet. Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.925347 4767 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.926379 4767 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.926410 4767 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.926540 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.929904 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.929936 4767 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.929991 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 21:00:54.111714417 +0000 UTC Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.930278 4767 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 29 15:00:44 crc kubenswrapper[4767]: E0129 15:00:44.930303 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.930317 4767 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.930339 4767 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 29 15:00:44 crc kubenswrapper[4767]: W0129 15:00:44.940807 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:44 crc kubenswrapper[4767]: E0129 15:00:44.940972 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.942556 4767 factory.go:153] Registering CRI-O factory Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.942647 4767 factory.go:221] Registration of the crio container factory successfully Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.942791 4767 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.942816 4767 factory.go:55] Registering systemd factory Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.942827 4767 factory.go:221] Registration of the systemd container factory successfully Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.942862 4767 factory.go:103] Registering Raw factory Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.942887 4767 manager.go:1196] Started watching for new ooms in manager Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.947329 4767 server.go:460] "Adding debug handlers to kubelet server" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.948081 4767 manager.go:319] Starting recovery of all containers Jan 29 15:00:44 crc kubenswrapper[4767]: E0129 15:00:44.948580 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188f3bb923122e54 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 15:00:44.923276884 +0000 UTC m=+0.733593943,LastTimestamp:2026-01-29 15:00:44.923276884 +0000 UTC m=+0.733593943,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 15:00:44 crc kubenswrapper[4767]: E0129 15:00:44.954461 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="200ms" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959127 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959173 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959183 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959193 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959204 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959213 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959222 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959231 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959242 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959252 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959261 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959270 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959279 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959289 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959298 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959307 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959316 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959323 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959331 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959341 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959353 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959368 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959378 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959387 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959396 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959406 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959419 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959429 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959437 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959445 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959452 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959461 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959470 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959478 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959487 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959496 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959505 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959514 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959522 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959533 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959541 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959552 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959563 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959575 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959588 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959599 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959611 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959624 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959655 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959664 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959674 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959683 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959696 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959706 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959717 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959728 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959737 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959745 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959758 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959767 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959778 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959787 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959795 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959805 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959814 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959824 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959832 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959841 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959849 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959859 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959867 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959876 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959885 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959909 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959920 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959928 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959937 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959945 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959955 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959965 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959973 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959982 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.959991 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960000 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960009 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960018 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960028 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960037 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960067 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960077 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960088 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960097 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960105 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960114 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960122 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960132 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960141 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960152 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960160 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960170 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960179 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960188 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960202 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960210 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.960223 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964176 4767 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964208 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964223 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964234 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964245 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964257 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964268 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964277 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964286 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964295 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964308 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964320 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964332 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964344 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964354 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964363 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964372 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964380 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964389 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964398 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964408 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964418 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964427 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964437 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964446 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964455 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964464 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964474 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964482 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964492 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964500 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964509 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964517 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964525 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964533 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964541 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964549 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964557 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964566 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964576 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964586 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964594 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964602 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964610 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964619 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964629 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964637 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964644 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964655 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964664 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964674 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964683 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964691 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964703 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964712 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964721 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964730 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964739 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964748 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964758 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964766 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964774 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964783 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964792 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964801 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964814 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964823 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964831 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964840 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964852 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964861 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964870 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964880 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964888 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964914 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964925 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964937 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964948 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964959 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964972 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964982 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.964993 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965003 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965011 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965026 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965037 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965046 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965055 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965065 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965074 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965083 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965092 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965101 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965111 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965119 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965126 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965135 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965145 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965154 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965164 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965175 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965185 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965195 4767 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965206 4767 reconstruct.go:97] "Volume reconstruction finished" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.965214 4767 reconciler.go:26] "Reconciler: start to sync state" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.975615 4767 manager.go:324] Recovery completed Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.986553 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.988238 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.988286 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.988299 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.989176 4767 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.989202 4767 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 29 15:00:44 crc kubenswrapper[4767]: I0129 15:00:44.989224 4767 state_mem.go:36] "Initialized new in-memory state store" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.012548 4767 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.017194 4767 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.017254 4767 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.017293 4767 kubelet.go:2335] "Starting kubelet main sync loop" Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.017443 4767 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 29 15:00:45 crc kubenswrapper[4767]: W0129 15:00:45.021866 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.021960 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.030911 4767 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.034376 4767 policy_none.go:49] "None policy: Start" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.035529 4767 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.035577 4767 state_mem.go:35] "Initializing new in-memory state store" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.096219 4767 manager.go:334] "Starting Device Plugin manager" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.097194 4767 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.097214 4767 server.go:79] "Starting device plugin registration server" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.097603 4767 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.097632 4767 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.097795 4767 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.097875 4767 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.097887 4767 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.105386 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.117579 4767 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.117681 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.118752 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.118788 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.118827 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.119011 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.119414 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.119474 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.119940 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.119970 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.119979 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120071 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120268 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120329 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120479 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120516 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120639 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120658 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120668 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120747 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120877 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.120944 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121243 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121279 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121367 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121390 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121399 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121466 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121492 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121511 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121840 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.121870 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.122532 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.122582 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.122592 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.122728 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.122750 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.123244 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.123261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.123269 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.123728 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.123751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.123762 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.155847 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="400ms" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167141 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167244 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167295 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167314 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167412 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167616 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167677 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167713 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167742 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167763 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167790 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167820 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167840 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167877 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.167915 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.198391 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.199544 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.199585 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.199597 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.199631 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.200259 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269607 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269673 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269708 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269737 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269766 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269796 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269827 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269856 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269886 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269941 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.269969 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270001 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270032 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270060 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270055 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270102 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270150 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270089 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270194 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270236 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270203 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270234 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270278 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270304 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270330 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270343 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270355 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270132 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270393 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.270402 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.401561 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.403204 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.403272 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.403289 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.403331 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.403970 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.458950 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.468141 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.487662 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.505450 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.512067 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 29 15:00:45 crc kubenswrapper[4767]: W0129 15:00:45.520147 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-25ceb84648d2f660c4338ceec180b8af6cbef7ccc682370e2758c72746bdd5c7 WatchSource:0}: Error finding container 25ceb84648d2f660c4338ceec180b8af6cbef7ccc682370e2758c72746bdd5c7: Status 404 returned error can't find the container with id 25ceb84648d2f660c4338ceec180b8af6cbef7ccc682370e2758c72746bdd5c7 Jan 29 15:00:45 crc kubenswrapper[4767]: W0129 15:00:45.524445 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-4c321a897a18a7fbbb4150942ee5a3a32d3fd4d2d5fb98236477fd02beb9b7cf WatchSource:0}: Error finding container 4c321a897a18a7fbbb4150942ee5a3a32d3fd4d2d5fb98236477fd02beb9b7cf: Status 404 returned error can't find the container with id 4c321a897a18a7fbbb4150942ee5a3a32d3fd4d2d5fb98236477fd02beb9b7cf Jan 29 15:00:45 crc kubenswrapper[4767]: W0129 15:00:45.538180 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-7ca2d8329dd79f29c5ca7a9f82e791a729e534650bdc77e30afab2bfcd71fec6 WatchSource:0}: Error finding container 7ca2d8329dd79f29c5ca7a9f82e791a729e534650bdc77e30afab2bfcd71fec6: Status 404 returned error can't find the container with id 7ca2d8329dd79f29c5ca7a9f82e791a729e534650bdc77e30afab2bfcd71fec6 Jan 29 15:00:45 crc kubenswrapper[4767]: W0129 15:00:45.539999 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-06eacc5ae8c5a67176c70b99fb0fabe32a4cee883562bf3e8fdfc098469a7de3 WatchSource:0}: Error finding container 06eacc5ae8c5a67176c70b99fb0fabe32a4cee883562bf3e8fdfc098469a7de3: Status 404 returned error can't find the container with id 06eacc5ae8c5a67176c70b99fb0fabe32a4cee883562bf3e8fdfc098469a7de3 Jan 29 15:00:45 crc kubenswrapper[4767]: W0129 15:00:45.546318 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-d586c60f07cc463bfd779370a0ab4f83d56a9ab01865e3c50cfb21015a059527 WatchSource:0}: Error finding container d586c60f07cc463bfd779370a0ab4f83d56a9ab01865e3c50cfb21015a059527: Status 404 returned error can't find the container with id d586c60f07cc463bfd779370a0ab4f83d56a9ab01865e3c50cfb21015a059527 Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.557700 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="800ms" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.804756 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.806266 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.806308 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.806323 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.806356 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.806788 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.927961 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:45 crc kubenswrapper[4767]: I0129 15:00:45.930983 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 08:41:24.885995761 +0000 UTC Jan 29 15:00:45 crc kubenswrapper[4767]: W0129 15:00:45.959027 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:45 crc kubenswrapper[4767]: E0129 15:00:45.959131 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.028311 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d586c60f07cc463bfd779370a0ab4f83d56a9ab01865e3c50cfb21015a059527"} Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.030356 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"06eacc5ae8c5a67176c70b99fb0fabe32a4cee883562bf3e8fdfc098469a7de3"} Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.031399 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7ca2d8329dd79f29c5ca7a9f82e791a729e534650bdc77e30afab2bfcd71fec6"} Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.033379 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"25ceb84648d2f660c4338ceec180b8af6cbef7ccc682370e2758c72746bdd5c7"} Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.035941 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4c321a897a18a7fbbb4150942ee5a3a32d3fd4d2d5fb98236477fd02beb9b7cf"} Jan 29 15:00:46 crc kubenswrapper[4767]: W0129 15:00:46.082217 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:46 crc kubenswrapper[4767]: E0129 15:00:46.082462 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:46 crc kubenswrapper[4767]: W0129 15:00:46.292519 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:46 crc kubenswrapper[4767]: E0129 15:00:46.292621 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:46 crc kubenswrapper[4767]: E0129 15:00:46.358890 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="1.6s" Jan 29 15:00:46 crc kubenswrapper[4767]: W0129 15:00:46.587047 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:46 crc kubenswrapper[4767]: E0129 15:00:46.587185 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.606993 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.608621 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.608665 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.608681 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.608714 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 15:00:46 crc kubenswrapper[4767]: E0129 15:00:46.609148 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.798594 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 15:00:46 crc kubenswrapper[4767]: E0129 15:00:46.801542 4767 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.927151 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:46 crc kubenswrapper[4767]: I0129 15:00:46.931355 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 07:24:09.108475624 +0000 UTC Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.040999 4767 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0" exitCode=0 Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.041111 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0"} Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.041213 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.042796 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.042944 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.042966 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.044470 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b"} Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.044528 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace"} Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.046935 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0" exitCode=0 Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.047006 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0"} Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.047110 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.048838 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.048944 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.048972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.049400 4767 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91" exitCode=0 Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.049610 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.049806 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91"} Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.050796 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.050858 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.050870 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.050881 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.052080 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.052131 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.052147 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.053186 4767 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c" exitCode=0 Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.053232 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c"} Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.053396 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.054653 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.054681 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.054698 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.927456 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:47 crc kubenswrapper[4767]: I0129 15:00:47.931509 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 08:09:29.178461875 +0000 UTC Jan 29 15:00:47 crc kubenswrapper[4767]: E0129 15:00:47.960134 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="3.2s" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.062271 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.062358 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.062301 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.062389 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.063480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.063534 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.063553 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.065996 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.066037 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.066150 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.067392 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.067680 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.067732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.069376 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.069427 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.069441 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.069450 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.071080 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d07aa7ed0ef0aa928fdd9985e1452a58f70b50129ce206e2e423a98b691f841e"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.071155 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.072037 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.072060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.072070 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.072640 4767 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e" exitCode=0 Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.072687 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e"} Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.072724 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.073595 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.073641 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.073657 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.209795 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.210861 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.210919 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.210935 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.210961 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 15:00:48 crc kubenswrapper[4767]: E0129 15:00:48.211452 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.138:6443: connect: connection refused" node="crc" Jan 29 15:00:48 crc kubenswrapper[4767]: W0129 15:00:48.450343 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:48 crc kubenswrapper[4767]: E0129 15:00:48.450426 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:48 crc kubenswrapper[4767]: W0129 15:00:48.473386 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:48 crc kubenswrapper[4767]: E0129 15:00:48.473472 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:48 crc kubenswrapper[4767]: W0129 15:00:48.486517 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:48 crc kubenswrapper[4767]: E0129 15:00:48.486591 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.927705 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:48 crc kubenswrapper[4767]: I0129 15:00:48.931927 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 21:09:49.814702774 +0000 UTC Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.077642 4767 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63" exitCode=0 Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.077759 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63"} Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.077802 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.078923 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.078952 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.078962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.080880 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"52a72166fa63a7a77d17247d32ea481f61dae4b419a4479e5b9535bc6be90934"} Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.080941 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.080984 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.080992 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.081015 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.081037 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082321 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082371 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082412 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082416 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082419 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.082429 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:49 crc kubenswrapper[4767]: W0129 15:00:49.213260 4767 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.138:6443: connect: connection refused Jan 29 15:00:49 crc kubenswrapper[4767]: E0129 15:00:49.213346 4767 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.138:6443: connect: connection refused" logger="UnhandledError" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.669455 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:49 crc kubenswrapper[4767]: I0129 15:00:49.933126 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 08:27:50.748232951 +0000 UTC Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.094577 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095010 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e"} Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095048 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c"} Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095062 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953"} Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095072 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103"} Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095080 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10"} Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095512 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095540 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.095549 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.097376 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.099164 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="52a72166fa63a7a77d17247d32ea481f61dae4b419a4479e5b9535bc6be90934" exitCode=255 Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.099203 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"52a72166fa63a7a77d17247d32ea481f61dae4b419a4479e5b9535bc6be90934"} Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.099288 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.100201 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.100219 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.100228 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.100755 4767 scope.go:117] "RemoveContainer" containerID="52a72166fa63a7a77d17247d32ea481f61dae4b419a4479e5b9535bc6be90934" Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.900510 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 29 15:00:50 crc kubenswrapper[4767]: I0129 15:00:50.933474 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 06:17:12.894023496 +0000 UTC Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.103701 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.105250 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.105704 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.105733 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.106059 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9"} Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.106498 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.106524 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.106532 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.107111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.107132 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.107141 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.305681 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.411570 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.413453 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.413536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.413563 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.413601 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.587727 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.588023 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.589735 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.589789 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.589802 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:51 crc kubenswrapper[4767]: I0129 15:00:51.934329 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 04:37:53.642605251 +0000 UTC Jan 29 15:00:52 crc kubenswrapper[4767]: I0129 15:00:52.108318 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 15:00:52 crc kubenswrapper[4767]: I0129 15:00:52.108389 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:52 crc kubenswrapper[4767]: I0129 15:00:52.109878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:52 crc kubenswrapper[4767]: I0129 15:00:52.109956 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:52 crc kubenswrapper[4767]: I0129 15:00:52.109974 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:52 crc kubenswrapper[4767]: I0129 15:00:52.578919 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:00:52 crc kubenswrapper[4767]: I0129 15:00:52.935292 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 21:55:56.391083411 +0000 UTC Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.111110 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.112360 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.112404 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.112421 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.296421 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.296659 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.298281 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.298344 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.298363 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.404461 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.404721 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.406362 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.406432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.406451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.889160 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.889421 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.890827 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.890876 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.890923 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:53 crc kubenswrapper[4767]: I0129 15:00:53.935747 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 17:14:14.382967667 +0000 UTC Jan 29 15:00:54 crc kubenswrapper[4767]: I0129 15:00:54.114371 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:54 crc kubenswrapper[4767]: I0129 15:00:54.115871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:54 crc kubenswrapper[4767]: I0129 15:00:54.115953 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:54 crc kubenswrapper[4767]: I0129 15:00:54.115969 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:54 crc kubenswrapper[4767]: I0129 15:00:54.936616 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 02:12:46.17440646 +0000 UTC Jan 29 15:00:55 crc kubenswrapper[4767]: E0129 15:00:55.105579 4767 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 29 15:00:55 crc kubenswrapper[4767]: I0129 15:00:55.796460 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:55 crc kubenswrapper[4767]: I0129 15:00:55.796648 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:55 crc kubenswrapper[4767]: I0129 15:00:55.797931 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:55 crc kubenswrapper[4767]: I0129 15:00:55.797962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:55 crc kubenswrapper[4767]: I0129 15:00:55.797972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:55 crc kubenswrapper[4767]: I0129 15:00:55.936964 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 12:41:58.206045635 +0000 UTC Jan 29 15:00:56 crc kubenswrapper[4767]: I0129 15:00:56.937542 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 01:45:13.558619818 +0000 UTC Jan 29 15:00:57 crc kubenswrapper[4767]: I0129 15:00:57.938033 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 07:44:10.772292238 +0000 UTC Jan 29 15:00:57 crc kubenswrapper[4767]: I0129 15:00:57.964792 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 29 15:00:57 crc kubenswrapper[4767]: I0129 15:00:57.965105 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:57 crc kubenswrapper[4767]: I0129 15:00:57.966954 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:57 crc kubenswrapper[4767]: I0129 15:00:57.967011 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:57 crc kubenswrapper[4767]: I0129 15:00:57.967030 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.001645 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.001852 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.003504 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.003570 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.003593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.010201 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.124277 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.126752 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.127012 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.127278 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.130290 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.798607 4767 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.798752 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 15:00:58 crc kubenswrapper[4767]: I0129 15:00:58.938878 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 17:49:36.2671589 +0000 UTC Jan 29 15:00:59 crc kubenswrapper[4767]: I0129 15:00:59.127035 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:00:59 crc kubenswrapper[4767]: I0129 15:00:59.128237 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:00:59 crc kubenswrapper[4767]: I0129 15:00:59.128283 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:00:59 crc kubenswrapper[4767]: I0129 15:00:59.128300 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:00:59 crc kubenswrapper[4767]: I0129 15:00:59.929943 4767 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 29 15:00:59 crc kubenswrapper[4767]: I0129 15:00:59.938999 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 19:27:01.072721986 +0000 UTC Jan 29 15:01:00 crc kubenswrapper[4767]: I0129 15:01:00.047295 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Jan 29 15:01:00 crc kubenswrapper[4767]: I0129 15:01:00.047398 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 29 15:01:00 crc kubenswrapper[4767]: I0129 15:01:00.053613 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Jan 29 15:01:00 crc kubenswrapper[4767]: I0129 15:01:00.053812 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 29 15:01:00 crc kubenswrapper[4767]: I0129 15:01:00.939940 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 19:00:36.571312881 +0000 UTC Jan 29 15:01:01 crc kubenswrapper[4767]: I0129 15:01:01.311011 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]log ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]etcd ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-startkubeinformers ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-api-request-count-filter ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/generic-apiserver-start-informers ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/priority-and-fairness-config-consumer ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/priority-and-fairness-filter ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-apiextensions-informers ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-apiextensions-controllers ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/crd-informer-synced ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-system-namespaces-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-cluster-authentication-info-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-legacy-token-tracking-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-service-ip-repair-controllers ok Jan 29 15:01:01 crc kubenswrapper[4767]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/priority-and-fairness-config-producer ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/bootstrap-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/start-kube-aggregator-informers ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/apiservice-status-local-available-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/apiservice-status-remote-available-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/apiservice-registration-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/apiservice-wait-for-first-sync ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/apiservice-discovery-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/kube-apiserver-autoregistration ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]autoregister-completion ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/apiservice-openapi-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: [+]poststarthook/apiservice-openapiv3-controller ok Jan 29 15:01:01 crc kubenswrapper[4767]: livez check failed Jan 29 15:01:01 crc kubenswrapper[4767]: I0129 15:01:01.311102 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:01:01 crc kubenswrapper[4767]: I0129 15:01:01.940873 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 20:44:02.82089582 +0000 UTC Jan 29 15:01:02 crc kubenswrapper[4767]: I0129 15:01:02.580261 4767 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 29 15:01:02 crc kubenswrapper[4767]: I0129 15:01:02.580345 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 29 15:01:02 crc kubenswrapper[4767]: I0129 15:01:02.941248 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 20:53:05.576861172 +0000 UTC Jan 29 15:01:03 crc kubenswrapper[4767]: I0129 15:01:03.942413 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 18:16:59.352084422 +0000 UTC Jan 29 15:01:04 crc kubenswrapper[4767]: I0129 15:01:04.942667 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 05:41:09.941403679 +0000 UTC Jan 29 15:01:05 crc kubenswrapper[4767]: E0129 15:01:05.019545 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.021814 4767 trace.go:236] Trace[777521073]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 15:00:54.467) (total time: 10553ms): Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[777521073]: ---"Objects listed" error: 10553ms (15:01:05.021) Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[777521073]: [10.553809628s] [10.553809628s] END Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.021862 4767 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.023686 4767 trace.go:236] Trace[1316764142]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 15:00:54.324) (total time: 10698ms): Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[1316764142]: ---"Objects listed" error: 10698ms (15:01:05.023) Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[1316764142]: [10.698894074s] [10.698894074s] END Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.023722 4767 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 29 15:01:05 crc kubenswrapper[4767]: E0129 15:01:05.024951 4767 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.025071 4767 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.025217 4767 trace.go:236] Trace[392375250]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 15:00:52.736) (total time: 12288ms): Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[392375250]: ---"Objects listed" error: 12288ms (15:01:05.025) Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[392375250]: [12.288551261s] [12.288551261s] END Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.025255 4767 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.026773 4767 trace.go:236] Trace[83053659]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (29-Jan-2026 15:00:53.453) (total time: 11573ms): Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[83053659]: ---"Objects listed" error: 11572ms (15:01:05.026) Jan 29 15:01:05 crc kubenswrapper[4767]: Trace[83053659]: [11.573217284s] [11.573217284s] END Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.026818 4767 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.063278 4767 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.801192 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.805302 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.921334 4767 apiserver.go:52] "Watching apiserver" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.924832 4767 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.925273 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.925863 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.926025 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.926035 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.926721 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:05 crc kubenswrapper[4767]: E0129 15:01:05.926734 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:05 crc kubenswrapper[4767]: E0129 15:01:05.926940 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.926989 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.927041 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:05 crc kubenswrapper[4767]: E0129 15:01:05.927189 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.929646 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.931202 4767 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.932333 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.932602 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.932614 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.933214 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.934088 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.935065 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.935224 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.935327 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.943259 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 03:19:17.582888911 +0000 UTC Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.958108 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:05 crc kubenswrapper[4767]: I0129 15:01:05.974094 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.008159 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.020605 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.030488 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.030991 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.031141 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.031244 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.031434 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.031543 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.031653 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.031753 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.031829 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032016 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032049 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032123 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032136 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032399 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032533 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032639 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032738 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032935 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033096 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033261 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032230 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032532 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032772 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.032965 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033003 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033172 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033238 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033308 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033362 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033361 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033579 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.034981 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035005 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035024 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035044 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035060 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035076 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035092 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035108 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035145 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035162 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035191 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035210 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035228 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035243 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035258 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035274 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035290 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035307 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035323 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035328 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035339 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035399 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035407 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035485 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035516 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035525 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035562 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035596 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035632 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035666 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035676 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035825 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035866 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035926 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035956 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.036520 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.036531 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.036527 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.036973 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033446 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033502 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033611 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033637 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.033876 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.036235 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.036338 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.037654 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.038009 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.038361 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.038632 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.039143 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.039556 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.040026 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.040038 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.040584 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.040849 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.035965 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.040747 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.036065 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:01:06.536047005 +0000 UTC m=+22.346364044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041068 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041252 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041378 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041387 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041427 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041459 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041484 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041479 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041519 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041551 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041577 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041605 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041634 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041661 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041685 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041710 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041738 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041761 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041789 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041848 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041885 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041942 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041955 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041999 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041965 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042170 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042184 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042253 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042286 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042310 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042336 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042361 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042383 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042424 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042475 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042474 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042549 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042571 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042593 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042617 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042650 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042670 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042691 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042708 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042728 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042748 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042768 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042774 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042792 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042881 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042914 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042944 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.042989 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043032 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043163 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043197 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043226 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043258 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043285 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043754 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043804 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044657 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043027 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044708 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043044 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043351 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044804 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044844 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044881 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044768 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.043399 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044033 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044058 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044335 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044392 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044644 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044647 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044855 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044873 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044887 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044946 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.044932 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045205 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045249 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045283 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045312 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045461 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045490 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045418 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045562 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045807 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045853 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045906 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045951 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.041334 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.046285 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.046281 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.046443 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.046517 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.046620 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.046681 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.047485 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.047578 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.045994 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.047633 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.047657 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.047735 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.047805 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.047829 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048070 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048088 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048171 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048183 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048196 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048236 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048372 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048427 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048458 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048491 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048592 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048484 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048723 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048764 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048793 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048821 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.048849 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049295 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049441 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049060 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049690 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049738 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049753 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049967 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049971 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050008 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050031 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050324 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050003 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050382 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.049868 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050365 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050408 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050506 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050529 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050551 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050568 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050595 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050599 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050614 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050686 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050803 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050833 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050890 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050954 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050991 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051030 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051066 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051102 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051134 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051167 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051202 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051231 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051264 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051295 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051331 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051363 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051393 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051426 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051458 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051490 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051521 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051551 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051585 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051619 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051643 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051667 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051692 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051740 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051761 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051784 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051807 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051831 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051853 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051874 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051918 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051942 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051964 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051987 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052012 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052043 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052075 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052101 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052128 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052161 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052192 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052216 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052282 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052316 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052349 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052382 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052421 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052456 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052489 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052522 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052554 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052588 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052624 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052659 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052691 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052724 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052759 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052793 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052829 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052862 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052919 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052964 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053030 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053074 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053108 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053157 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053188 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053212 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053237 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053262 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053297 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053331 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053364 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053402 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053441 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053468 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053557 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053582 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.050991 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053602 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053622 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053641 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053661 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053680 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053700 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053717 4767 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053733 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053752 4767 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053773 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053791 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053809 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053829 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053846 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053862 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053879 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053934 4767 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053954 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053971 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053989 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054007 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054024 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054042 4767 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054059 4767 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054077 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054094 4767 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054111 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054130 4767 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054151 4767 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054173 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054192 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054237 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054256 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054276 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054294 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054313 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054330 4767 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054348 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054366 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054385 4767 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054403 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054419 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054436 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054454 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054470 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054487 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054506 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054523 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054544 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054563 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054580 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054596 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054613 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054632 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054648 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054665 4767 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054681 4767 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054699 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054717 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054733 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054750 4767 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054790 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054805 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054821 4767 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054838 4767 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054858 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054876 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054917 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054938 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054955 4767 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054972 4767 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054990 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055008 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055025 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055044 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055061 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055078 4767 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055097 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055177 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055194 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055212 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055229 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055246 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055263 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055285 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055300 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055318 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055336 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055355 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055372 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055389 4767 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055406 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055424 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055443 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055460 4767 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055480 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055500 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055519 4767 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055537 4767 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055557 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055573 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055592 4767 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055611 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055629 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055646 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055664 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055682 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053602 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051072 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051113 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051249 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051258 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.046144 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051450 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051556 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051691 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051754 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.051953 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052059 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052169 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052219 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052439 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052448 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.052858 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053094 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053219 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053367 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053585 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.053799 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054032 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054062 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054282 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.054432 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.055410 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.056386 4767 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.057215 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.057478 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.057572 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.057758 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.057920 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.057977 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.057987 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.058436 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.058487 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.058835 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.058984 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.059025 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.059095 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.059280 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.059253 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.059643 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.059884 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.060232 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.060389 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.060542 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.060601 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.060733 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.060983 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.061052 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.061283 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.061293 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.061327 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.061489 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.061775 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.061888 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.062252 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.062294 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.062312 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.062473 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.063373 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.063868 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.063986 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:06.563960045 +0000 UTC m=+22.374277284 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.065241 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.065381 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.065408 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.065445 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.065498 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.065528 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:06.565501951 +0000 UTC m=+22.375818990 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.066035 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.066719 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.076355 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.076390 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.076408 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.076493 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:06.576469063 +0000 UTC m=+22.386786102 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.080288 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.080488 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.080511 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.080526 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.080594 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:06.580573574 +0000 UTC m=+22.390890613 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.083469 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.083608 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.083788 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.084060 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.084076 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.085834 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.088950 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.088999 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.090593 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.090607 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.091887 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.113962 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114177 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114338 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114467 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114517 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114523 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114732 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114774 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114871 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.114977 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.118708 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.130138 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.132590 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.141198 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.143920 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.147834 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.148875 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.151381 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" exitCode=255 Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.151997 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9"} Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.152053 4767 scope.go:117] "RemoveContainer" containerID="52a72166fa63a7a77d17247d32ea481f61dae4b419a4479e5b9535bc6be90934" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159703 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159734 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159794 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159806 4767 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159820 4767 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159832 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159842 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159852 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159862 4767 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159872 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159881 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159904 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159914 4767 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159923 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159885 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159977 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.159935 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160349 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160371 4767 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160385 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160399 4767 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160411 4767 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160422 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160433 4767 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160445 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160458 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160470 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160481 4767 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160493 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160505 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160516 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160527 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160538 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160549 4767 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160560 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160572 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160585 4767 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160597 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160608 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160620 4767 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160632 4767 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160643 4767 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160655 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160667 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160678 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160696 4767 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160708 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160722 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160758 4767 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160770 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160782 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160793 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160804 4767 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160832 4767 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160844 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160857 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160869 4767 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160880 4767 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160967 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.160988 4767 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161001 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161013 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161024 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161035 4767 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161045 4767 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161057 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161068 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161083 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161096 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161107 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161121 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161131 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161142 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161153 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161164 4767 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161175 4767 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161185 4767 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161194 4767 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161204 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161214 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161223 4767 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161234 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161244 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161256 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161266 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161279 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161290 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161300 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.161312 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.163620 4767 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.167619 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.169848 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.170401 4767 scope.go:117] "RemoveContainer" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.170623 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.177725 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.186369 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.195600 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.203414 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.211648 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.218421 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.249924 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 29 15:01:06 crc kubenswrapper[4767]: W0129 15:01:06.260835 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-6ddbcf547365aa95d35d9f575b46a7429b56c3f7699bc20b407760a074502261 WatchSource:0}: Error finding container 6ddbcf547365aa95d35d9f575b46a7429b56c3f7699bc20b407760a074502261: Status 404 returned error can't find the container with id 6ddbcf547365aa95d35d9f575b46a7429b56c3f7699bc20b407760a074502261 Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.267568 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.278508 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 29 15:01:06 crc kubenswrapper[4767]: W0129 15:01:06.297146 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-0649ecbd6257cff0cc50ef45e798bda714d3593fd9a37bf122589f7810d4e76a WatchSource:0}: Error finding container 0649ecbd6257cff0cc50ef45e798bda714d3593fd9a37bf122589f7810d4e76a: Status 404 returned error can't find the container with id 0649ecbd6257cff0cc50ef45e798bda714d3593fd9a37bf122589f7810d4e76a Jan 29 15:01:06 crc kubenswrapper[4767]: W0129 15:01:06.300919 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3fa4e9519bdbfa9d22367f2aa1ea5014b15bfd181317596117882bd8fede0d26 WatchSource:0}: Error finding container 3fa4e9519bdbfa9d22367f2aa1ea5014b15bfd181317596117882bd8fede0d26: Status 404 returned error can't find the container with id 3fa4e9519bdbfa9d22367f2aa1ea5014b15bfd181317596117882bd8fede0d26 Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.311069 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.324686 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52a72166fa63a7a77d17247d32ea481f61dae4b419a4479e5b9535bc6be90934\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"message\\\":\\\"W0129 15:00:48.723451 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 15:00:48.724812 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769698848 cert, and key in /tmp/serving-cert-1136208431/serving-signer.crt, /tmp/serving-cert-1136208431/serving-signer.key\\\\nI0129 15:00:48.980778 1 observer_polling.go:159] Starting file observer\\\\nW0129 15:00:48.983601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 15:00:48.983713 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:00:48.985031 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1136208431/tls.crt::/tmp/serving-cert-1136208431/tls.key\\\\\\\"\\\\nF0129 15:00:49.594134 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.334071 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.344951 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.359938 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.386964 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.399593 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.413385 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.422178 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.563929 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.564026 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.564126 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:01:07.564103077 +0000 UTC m=+23.374420126 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.564177 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.564239 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:07.564225681 +0000 UTC m=+23.374542730 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.664385 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.664440 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.664475 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664574 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664589 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664600 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664643 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:07.664627942 +0000 UTC m=+23.474944981 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664644 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664747 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:07.664725785 +0000 UTC m=+23.475042854 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664874 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664913 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664928 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: E0129 15:01:06.664968 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:07.664959092 +0000 UTC m=+23.475276211 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:06 crc kubenswrapper[4767]: I0129 15:01:06.944263 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 02:52:39.52769509 +0000 UTC Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.018110 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.018355 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.026121 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.027316 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.029793 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.031709 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.034093 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.035105 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.036764 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.039073 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.041193 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.043939 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.045780 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.049747 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.052067 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.053669 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.056367 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.057938 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.060269 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.061451 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.063153 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.066269 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.068228 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.070606 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.071735 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.073522 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.074393 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.075245 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.076331 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.077686 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.079292 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.080533 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.082679 4767 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.082933 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.087036 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.088492 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.089154 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.090295 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.091031 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.091568 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.092205 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.092839 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.093306 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.093874 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.094494 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.096280 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.096923 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.098300 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.098912 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.100237 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.100800 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.101321 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.102349 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.102940 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.104052 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.104504 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.159446 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3fa4e9519bdbfa9d22367f2aa1ea5014b15bfd181317596117882bd8fede0d26"} Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.161244 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07"} Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.161285 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59"} Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.161299 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0649ecbd6257cff0cc50ef45e798bda714d3593fd9a37bf122589f7810d4e76a"} Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.165033 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa"} Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.165077 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6ddbcf547365aa95d35d9f575b46a7429b56c3f7699bc20b407760a074502261"} Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.168076 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.173132 4767 scope.go:117] "RemoveContainer" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.173390 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.177814 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.179405 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.196881 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.209337 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.219653 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.232961 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52a72166fa63a7a77d17247d32ea481f61dae4b419a4479e5b9535bc6be90934\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"message\\\":\\\"W0129 15:00:48.723451 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0129 15:00:48.724812 1 crypto.go:601] Generating new CA for check-endpoints-signer@1769698848 cert, and key in /tmp/serving-cert-1136208431/serving-signer.crt, /tmp/serving-cert-1136208431/serving-signer.key\\\\nI0129 15:00:48.980778 1 observer_polling.go:159] Starting file observer\\\\nW0129 15:00:48.983601 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0129 15:00:48.983713 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:00:48.985031 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1136208431/tls.crt::/tmp/serving-cert-1136208431/tls.key\\\\\\\"\\\\nF0129 15:00:49.594134 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.248879 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.267052 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.278107 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.289116 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.298469 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.306662 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.316982 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.326769 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.337874 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.354432 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.366823 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.570465 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.570556 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.570645 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.570694 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:01:09.570656393 +0000 UTC m=+25.380973432 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.570733 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:09.570723075 +0000 UTC m=+25.381040264 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.671469 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.671525 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.671548 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.671626 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.671672 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:09.671659012 +0000 UTC m=+25.481976041 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672036 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672048 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672058 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672082 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:09.672075455 +0000 UTC m=+25.482392494 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672146 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672173 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672184 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:07 crc kubenswrapper[4767]: E0129 15:01:07.672233 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:09.672218319 +0000 UTC m=+25.482535358 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.944448 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 08:58:09.389759705 +0000 UTC Jan 29 15:01:07 crc kubenswrapper[4767]: I0129 15:01:07.991434 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.004541 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.009173 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.012640 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.017931 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.018024 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:08 crc kubenswrapper[4767]: E0129 15:01:08.018142 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:08 crc kubenswrapper[4767]: E0129 15:01:08.018269 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.019839 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.034833 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.071736 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.084008 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.095992 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.111451 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.124747 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.138675 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.153984 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.171991 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.173391 4767 scope.go:117] "RemoveContainer" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" Jan 29 15:01:08 crc kubenswrapper[4767]: E0129 15:01:08.173536 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.184808 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.198199 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.219784 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.235695 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.248702 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.262094 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:08Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:08 crc kubenswrapper[4767]: I0129 15:01:08.944687 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 05:59:38.049565954 +0000 UTC Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.018172 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.018299 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.177646 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea"} Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.178340 4767 scope.go:117] "RemoveContainer" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.179164 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.204002 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.222249 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.257288 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.272507 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.282411 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.300113 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.310065 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.320712 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.331808 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:09Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.363193 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.586083 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.586239 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:01:13.586212511 +0000 UTC m=+29.396529580 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.586313 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.586439 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.586500 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:13.586486499 +0000 UTC m=+29.396803548 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.687665 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.687719 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.687751 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687836 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687838 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687878 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687914 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687920 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687948 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687962 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.687932 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:13.68788632 +0000 UTC m=+29.498203379 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.688045 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:13.688015254 +0000 UTC m=+29.498332303 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:09 crc kubenswrapper[4767]: E0129 15:01:09.688068 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:13.688056055 +0000 UTC m=+29.498373194 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:09 crc kubenswrapper[4767]: I0129 15:01:09.945244 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 13:27:05.810782323 +0000 UTC Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.018152 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.018230 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:10 crc kubenswrapper[4767]: E0129 15:01:10.018279 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:10 crc kubenswrapper[4767]: E0129 15:01:10.018369 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.061278 4767 csr.go:261] certificate signing request csr-wg46f is approved, waiting to be issued Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.076649 4767 csr.go:257] certificate signing request csr-wg46f is issued Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.183937 4767 scope.go:117] "RemoveContainer" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" Jan 29 15:01:10 crc kubenswrapper[4767]: E0129 15:01:10.184143 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.946306 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 02:18:13.524407731 +0000 UTC Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.966071 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-k6xcf"] Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.966405 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.967284 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-9xcr9"] Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.967601 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.968788 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.968891 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.969152 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.969287 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.969670 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tg5zk"] Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.970011 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.970204 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.970351 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-kgsln"] Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.970535 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.970564 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-2qt9f"] Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.970633 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.970547 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.971208 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.971499 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.972382 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.972482 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.972680 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.972756 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.972947 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.974188 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.976098 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.976098 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.976107 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.976337 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.976336 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.976393 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.976794 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.978563 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.987268 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:10Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.997704 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4615231a-e157-430e-9ecc-97c3fd7701bb-cni-binary-copy\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.997927 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-netns\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998030 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-socket-dir-parent\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998106 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-conf-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998181 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcp9z\" (UniqueName: \"kubernetes.io/projected/4615231a-e157-430e-9ecc-97c3fd7701bb-kube-api-access-mcp9z\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998270 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-system-cni-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998342 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-cni-bin\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998418 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98h48\" (UniqueName: \"kubernetes.io/projected/afd48dc4-70ed-45d0-8133-5a6f2ef7014f-kube-api-access-98h48\") pod \"node-resolver-k6xcf\" (UID: \"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\") " pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998498 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-cnibin\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998600 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-daemon-config\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998641 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-multus-certs\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998682 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-k8s-cni-cncf-io\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998701 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-hostroot\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998719 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/afd48dc4-70ed-45d0-8133-5a6f2ef7014f-hosts-file\") pod \"node-resolver-k6xcf\" (UID: \"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\") " pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998741 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-os-release\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998756 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-etc-kubernetes\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998816 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-cni-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998841 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-kubelet\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:10 crc kubenswrapper[4767]: I0129 15:01:10.998858 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-cni-multus\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.007333 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.017565 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:11 crc kubenswrapper[4767]: E0129 15:01:11.017665 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.030038 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.044449 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.066857 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.077564 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-29 14:56:10 +0000 UTC, rotation deadline is 2026-10-19 15:42:55.770654225 +0000 UTC Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.077627 4767 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6312h41m44.693030024s for next certificate rotation Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.093750 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099375 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-env-overrides\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099418 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c144460d-d956-4e9c-8354-bfd72b970e30-mcd-auth-proxy-config\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099446 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cni-binary-copy\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099483 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98h48\" (UniqueName: \"kubernetes.io/projected/afd48dc4-70ed-45d0-8133-5a6f2ef7014f-kube-api-access-98h48\") pod \"node-resolver-k6xcf\" (UID: \"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\") " pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099506 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-kubelet\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099528 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-slash\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099551 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099572 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-os-release\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099592 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj6dz\" (UniqueName: \"kubernetes.io/projected/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-kube-api-access-cj6dz\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099615 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-cnibin\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099640 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-hostroot\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099660 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/afd48dc4-70ed-45d0-8133-5a6f2ef7014f-hosts-file\") pod \"node-resolver-k6xcf\" (UID: \"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\") " pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099681 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-systemd-units\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099700 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099719 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-ovn\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099741 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-system-cni-dir\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099761 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw776\" (UniqueName: \"kubernetes.io/projected/633e55cd-6740-4d13-84ef-023b691c0428-kube-api-access-dw776\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099782 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-os-release\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099804 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-etc-kubernetes\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099824 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-systemd\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099822 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-hostroot\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099844 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-kubelet\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099868 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099891 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-cni-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099931 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-cni-multus\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099956 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-var-lib-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099978 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-etc-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100001 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c144460d-d956-4e9c-8354-bfd72b970e30-rootfs\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100024 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-netns\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100044 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-script-lib\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100066 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4615231a-e157-430e-9ecc-97c3fd7701bb-cni-binary-copy\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100098 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-socket-dir-parent\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100119 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-conf-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100140 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcp9z\" (UniqueName: \"kubernetes.io/projected/4615231a-e157-430e-9ecc-97c3fd7701bb-kube-api-access-mcp9z\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100162 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-config\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100195 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-system-cni-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100219 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-cni-bin\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100236 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-os-release\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100240 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/633e55cd-6740-4d13-84ef-023b691c0428-ovn-node-metrics-cert\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100489 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100504 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-etc-kubernetes\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100533 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-daemon-config\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100555 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-multus-certs\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100573 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-ovn-kubernetes\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100593 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-bin\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100607 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c144460d-d956-4e9c-8354-bfd72b970e30-proxy-tls\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100622 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5hbb\" (UniqueName: \"kubernetes.io/projected/c144460d-d956-4e9c-8354-bfd72b970e30-kube-api-access-n5hbb\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100636 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-netd\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100653 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-k8s-cni-cncf-io\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100669 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-netns\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100685 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-log-socket\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100707 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-node-log\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100724 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cnibin\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100730 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-cni-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.099937 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-cnibin\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100775 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-cni-multus\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100823 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-netns\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101215 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-socket-dir-parent\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101284 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-conf-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101447 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4615231a-e157-430e-9ecc-97c3fd7701bb-multus-daemon-config\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101482 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-multus-certs\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.100554 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-kubelet\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101545 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-run-k8s-cni-cncf-io\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101589 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-host-var-lib-cni-bin\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101605 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4615231a-e157-430e-9ecc-97c3fd7701bb-cni-binary-copy\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101628 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/afd48dc4-70ed-45d0-8133-5a6f2ef7014f-hosts-file\") pod \"node-resolver-k6xcf\" (UID: \"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\") " pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.101633 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4615231a-e157-430e-9ecc-97c3fd7701bb-system-cni-dir\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.116552 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.126501 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcp9z\" (UniqueName: \"kubernetes.io/projected/4615231a-e157-430e-9ecc-97c3fd7701bb-kube-api-access-mcp9z\") pod \"multus-9xcr9\" (UID: \"4615231a-e157-430e-9ecc-97c3fd7701bb\") " pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.126501 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98h48\" (UniqueName: \"kubernetes.io/projected/afd48dc4-70ed-45d0-8133-5a6f2ef7014f-kube-api-access-98h48\") pod \"node-resolver-k6xcf\" (UID: \"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\") " pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.141380 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.156633 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.181093 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.191537 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.201948 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-config\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202253 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/633e55cd-6740-4d13-84ef-023b691c0428-ovn-node-metrics-cert\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202333 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202034 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202413 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-ovn-kubernetes\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202556 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-bin\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202586 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c144460d-d956-4e9c-8354-bfd72b970e30-proxy-tls\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202605 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5hbb\" (UniqueName: \"kubernetes.io/projected/c144460d-d956-4e9c-8354-bfd72b970e30-kube-api-access-n5hbb\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202623 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-netd\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202642 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-netns\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202657 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-log-socket\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202679 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-bin\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202760 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-netd\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202795 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-netns\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.202913 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-log-socket\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203041 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203043 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-ovn-kubernetes\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203168 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-node-log\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203193 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cnibin\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203212 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-env-overrides\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203230 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c144460d-d956-4e9c-8354-bfd72b970e30-mcd-auth-proxy-config\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203247 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-config\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203254 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cni-binary-copy\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203311 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-kubelet\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203332 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-slash\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203353 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203380 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-os-release\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203401 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj6dz\" (UniqueName: \"kubernetes.io/projected/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-kube-api-access-cj6dz\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203426 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-systemd-units\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203446 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203466 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-ovn\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203528 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cnibin\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203570 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-node-log\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203736 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-systemd-units\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203737 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-os-release\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203806 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-slash\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203815 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203845 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203879 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cni-binary-copy\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.203939 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-ovn\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204020 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-env-overrides\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204097 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-system-cni-dir\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204127 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw776\" (UniqueName: \"kubernetes.io/projected/633e55cd-6740-4d13-84ef-023b691c0428-kube-api-access-dw776\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204255 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-systemd\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204173 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-kubelet\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204329 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-systemd\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204362 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204184 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-system-cni-dir\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204387 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-var-lib-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-etc-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204467 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c144460d-d956-4e9c-8354-bfd72b970e30-rootfs\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204487 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-script-lib\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204524 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-etc-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.204565 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c144460d-d956-4e9c-8354-bfd72b970e30-rootfs\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.205060 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.205106 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-script-lib\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.205127 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-var-lib-openvswitch\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.205277 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/633e55cd-6740-4d13-84ef-023b691c0428-ovn-node-metrics-cert\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.205576 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c144460d-d956-4e9c-8354-bfd72b970e30-proxy-tls\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.205809 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c144460d-d956-4e9c-8354-bfd72b970e30-mcd-auth-proxy-config\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.222267 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj6dz\" (UniqueName: \"kubernetes.io/projected/65812f82-4e2a-4f3d-bf4c-3aa608f3de8d-kube-api-access-cj6dz\") pod \"multus-additional-cni-plugins-2qt9f\" (UID: \"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\") " pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.222298 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5hbb\" (UniqueName: \"kubernetes.io/projected/c144460d-d956-4e9c-8354-bfd72b970e30-kube-api-access-n5hbb\") pod \"machine-config-daemon-kgsln\" (UID: \"c144460d-d956-4e9c-8354-bfd72b970e30\") " pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.223383 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw776\" (UniqueName: \"kubernetes.io/projected/633e55cd-6740-4d13-84ef-023b691c0428-kube-api-access-dw776\") pod \"ovnkube-node-tg5zk\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.227631 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.251528 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.261497 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.273801 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.278669 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-k6xcf" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.285787 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.288091 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-9xcr9" Jan 29 15:01:11 crc kubenswrapper[4767]: W0129 15:01:11.288881 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafd48dc4_70ed_45d0_8133_5a6f2ef7014f.slice/crio-f11ba78e8d4d06d96cb7f2d86d128eec9b3f191063eee9daf951d761f34ef2d3 WatchSource:0}: Error finding container f11ba78e8d4d06d96cb7f2d86d128eec9b3f191063eee9daf951d761f34ef2d3: Status 404 returned error can't find the container with id f11ba78e8d4d06d96cb7f2d86d128eec9b3f191063eee9daf951d761f34ef2d3 Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.302377 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.302822 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.306792 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.316183 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.318435 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" Jan 29 15:01:11 crc kubenswrapper[4767]: W0129 15:01:11.323594 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod633e55cd_6740_4d13_84ef_023b691c0428.slice/crio-6660b45896ac7dd6baa4aa6e19ab9a27d30bbe942e7ba0aa3826282c494afe7b WatchSource:0}: Error finding container 6660b45896ac7dd6baa4aa6e19ab9a27d30bbe942e7ba0aa3826282c494afe7b: Status 404 returned error can't find the container with id 6660b45896ac7dd6baa4aa6e19ab9a27d30bbe942e7ba0aa3826282c494afe7b Jan 29 15:01:11 crc kubenswrapper[4767]: W0129 15:01:11.325173 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc144460d_d956_4e9c_8354_bfd72b970e30.slice/crio-12dedbad84c03bbf2feeb6a04a5f4e4bf5d0570fa7641bd3f9874247c7dfc2cd WatchSource:0}: Error finding container 12dedbad84c03bbf2feeb6a04a5f4e4bf5d0570fa7641bd3f9874247c7dfc2cd: Status 404 returned error can't find the container with id 12dedbad84c03bbf2feeb6a04a5f4e4bf5d0570fa7641bd3f9874247c7dfc2cd Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.327278 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.337945 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: W0129 15:01:11.345229 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65812f82_4e2a_4f3d_bf4c_3aa608f3de8d.slice/crio-e286ef0522be6e3ec68550742725f54c19fbef5094710ebb9f286c5e2eea5e4d WatchSource:0}: Error finding container e286ef0522be6e3ec68550742725f54c19fbef5094710ebb9f286c5e2eea5e4d: Status 404 returned error can't find the container with id e286ef0522be6e3ec68550742725f54c19fbef5094710ebb9f286c5e2eea5e4d Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.350483 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.367271 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.378406 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.425828 4767 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.432829 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.432877 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.432888 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.433018 4767 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.439677 4767 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.439784 4767 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.440955 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.441007 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.441016 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.441030 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.441038 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: E0129 15:01:11.463624 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.467866 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.467926 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.467940 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.467957 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.467969 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: E0129 15:01:11.480280 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.483692 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.483729 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.483738 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.483755 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.483764 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: E0129 15:01:11.495920 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.498726 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.498747 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.498756 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.498769 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.498777 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: E0129 15:01:11.512300 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.515996 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.516038 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.516050 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.516067 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.516077 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: E0129 15:01:11.531009 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:11Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:11 crc kubenswrapper[4767]: E0129 15:01:11.531118 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.537104 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.537194 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.537205 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.537224 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.537236 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.640078 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.640108 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.640118 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.640133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.640143 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.742784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.742817 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.742827 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.742844 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.742856 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.845165 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.845212 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.845225 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.845245 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.845256 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.946448 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 11:15:49.424877413 +0000 UTC Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.948997 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.949027 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.949036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.949048 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:11 crc kubenswrapper[4767]: I0129 15:01:11.949057 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:11Z","lastTransitionTime":"2026-01-29T15:01:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.018117 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.018145 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:12 crc kubenswrapper[4767]: E0129 15:01:12.018297 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:12 crc kubenswrapper[4767]: E0129 15:01:12.018444 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.051637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.051677 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.051686 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.051698 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.051707 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.154474 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.154516 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.154526 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.154541 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.154550 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.185952 4767 generic.go:334] "Generic (PLEG): container finished" podID="65812f82-4e2a-4f3d-bf4c-3aa608f3de8d" containerID="99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8" exitCode=0 Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.186012 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerDied","Data":"99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.186037 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerStarted","Data":"e286ef0522be6e3ec68550742725f54c19fbef5094710ebb9f286c5e2eea5e4d"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.188784 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.188816 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.188825 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"12dedbad84c03bbf2feeb6a04a5f4e4bf5d0570fa7641bd3f9874247c7dfc2cd"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.190756 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad" exitCode=0 Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.190818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.190839 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"6660b45896ac7dd6baa4aa6e19ab9a27d30bbe942e7ba0aa3826282c494afe7b"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.193287 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-k6xcf" event={"ID":"afd48dc4-70ed-45d0-8133-5a6f2ef7014f","Type":"ContainerStarted","Data":"caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.193325 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-k6xcf" event={"ID":"afd48dc4-70ed-45d0-8133-5a6f2ef7014f","Type":"ContainerStarted","Data":"f11ba78e8d4d06d96cb7f2d86d128eec9b3f191063eee9daf951d761f34ef2d3"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.194391 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerStarted","Data":"4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.194421 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerStarted","Data":"d0d61bae4c8a7f95b0f6525f5f2eeae112ee50477d5600734371506d61407083"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.204544 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.220398 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.230322 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.245454 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.257636 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.257674 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.257685 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.257701 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.257713 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.260798 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.271342 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.283217 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.297246 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.312579 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.326836 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.348865 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.360585 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.360623 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.360634 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.360649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.360660 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.360741 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.378254 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.400540 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.418096 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.440260 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.460813 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.463708 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.463783 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.463797 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.463816 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.463833 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.477447 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.489185 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.502875 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.516011 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.527818 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.537105 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.548876 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.561599 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.566195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.566242 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.566257 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.566276 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.566287 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.573147 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.584652 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.596139 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:12Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.668882 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.668939 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.668952 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.668968 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.668979 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.771455 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.771507 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.771519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.771537 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.771550 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.880278 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.880719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.880734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.880753 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.880776 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.946772 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 05:09:26.124747986 +0000 UTC Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.984282 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.984327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.984341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.984359 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:12 crc kubenswrapper[4767]: I0129 15:01:12.984371 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:12Z","lastTransitionTime":"2026-01-29T15:01:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.018370 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.018493 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.087656 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.087678 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.087686 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.087699 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.087708 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.190523 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.190934 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.190948 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.190965 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.190975 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.199735 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerStarted","Data":"c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.204569 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.204613 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.204624 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.204635 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.204647 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.223703 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.239042 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.252053 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.272462 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.285362 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.293211 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.293255 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.293265 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.293280 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.293292 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.297579 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.307468 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.322026 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.341405 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.352033 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.362395 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.372990 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.383284 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.392370 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.396168 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.396191 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.396199 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.396213 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.396223 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.498247 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.498281 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.498289 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.498302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.498312 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.581598 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zqvls"] Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.581985 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.584060 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.585129 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.585176 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.585245 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.600323 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.600373 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.600383 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.600428 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.600449 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.602698 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.615930 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.628749 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.628832 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3395d14-43f2-45f5-8692-034412aff69e-serviceca\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.628919 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:01:21.628862332 +0000 UTC m=+37.439179381 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.628978 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfv7b\" (UniqueName: \"kubernetes.io/projected/c3395d14-43f2-45f5-8692-034412aff69e-kube-api-access-bfv7b\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.629044 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3395d14-43f2-45f5-8692-034412aff69e-host\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.629098 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.629267 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.629329 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:21.629320395 +0000 UTC m=+37.439637434 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.629573 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.645383 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.656342 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.668813 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.677279 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.688392 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.701280 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.702478 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.702508 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.702519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.702535 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.702546 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.712187 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.722675 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.730237 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.730310 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfv7b\" (UniqueName: \"kubernetes.io/projected/c3395d14-43f2-45f5-8692-034412aff69e-kube-api-access-bfv7b\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.730333 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.730357 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3395d14-43f2-45f5-8692-034412aff69e-host\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.730399 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3395d14-43f2-45f5-8692-034412aff69e-serviceca\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730414 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730438 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.730445 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c3395d14-43f2-45f5-8692-034412aff69e-host\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730454 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730486 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730501 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730511 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730520 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:21.730503309 +0000 UTC m=+37.540820348 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730542 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:21.7305323 +0000 UTC m=+37.540849339 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.730422 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730803 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: E0129 15:01:13.730855 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:21.730844869 +0000 UTC m=+37.541161908 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.731360 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c3395d14-43f2-45f5-8692-034412aff69e-serviceca\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.736102 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.748029 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.749790 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfv7b\" (UniqueName: \"kubernetes.io/projected/c3395d14-43f2-45f5-8692-034412aff69e-kube-api-access-bfv7b\") pod \"node-ca-zqvls\" (UID: \"c3395d14-43f2-45f5-8692-034412aff69e\") " pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.758962 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.767616 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.804277 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.804315 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.804329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.804350 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.804364 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.893337 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zqvls" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.906003 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.906034 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.906047 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.906063 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.906093 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:13Z","lastTransitionTime":"2026-01-29T15:01:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:13 crc kubenswrapper[4767]: I0129 15:01:13.946959 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 04:47:20.308229444 +0000 UTC Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.008282 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.008304 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.008312 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.008324 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.008334 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.018156 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:14 crc kubenswrapper[4767]: E0129 15:01:14.018307 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.018802 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:14 crc kubenswrapper[4767]: E0129 15:01:14.018940 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.110220 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.110252 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.110260 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.110274 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.110282 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.212883 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.213007 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.213043 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.213058 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.213081 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.213097 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.220687 4767 generic.go:334] "Generic (PLEG): container finished" podID="65812f82-4e2a-4f3d-bf4c-3aa608f3de8d" containerID="c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826" exitCode=0 Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.220805 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerDied","Data":"c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.226513 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zqvls" event={"ID":"c3395d14-43f2-45f5-8692-034412aff69e","Type":"ContainerStarted","Data":"08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.226595 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zqvls" event={"ID":"c3395d14-43f2-45f5-8692-034412aff69e","Type":"ContainerStarted","Data":"6736c630ae0cca90a3062bbf9c55e61261eacbc44f4e0a187090a4c7d812fe09"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.244572 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.260474 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.269962 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.281936 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.297913 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.310742 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.316083 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.316124 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.316136 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.316152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.316164 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.323313 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.334924 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.347480 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.362449 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.375102 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.398120 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.411091 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.419519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.419562 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.419595 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.419616 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.419626 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.426718 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.446828 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.459412 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.471581 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.481564 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.492097 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.506558 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.517440 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.521710 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.521743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.521753 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.521767 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.521781 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.529105 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.540528 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.552671 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.562628 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.570947 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.586919 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.597080 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.609539 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.624266 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.624318 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.624339 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.624369 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.624390 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.625554 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:14Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.692838 4767 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.726719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.726754 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.726763 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.726777 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.726786 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.830170 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.830244 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.830268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.830302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.830324 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.932527 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.932592 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.932607 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.932633 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.932648 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:14Z","lastTransitionTime":"2026-01-29T15:01:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:14 crc kubenswrapper[4767]: I0129 15:01:14.947833 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 05:44:05.099728518 +0000 UTC Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.017783 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:15 crc kubenswrapper[4767]: E0129 15:01:15.017927 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.029564 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.034766 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.034795 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.034805 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.034818 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.034828 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.054723 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.074139 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.087571 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.097290 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.119211 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.134740 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.137866 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.137928 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.137941 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.137958 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.137970 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.149345 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.158884 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.172466 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.186587 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.198317 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.215642 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.232982 4767 generic.go:334] "Generic (PLEG): container finished" podID="65812f82-4e2a-4f3d-bf4c-3aa608f3de8d" containerID="a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94" exitCode=0 Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.233059 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerDied","Data":"a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.240168 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.240244 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.240270 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.240299 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.240327 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.241112 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.253200 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.264207 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.275689 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.298218 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.318881 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.329045 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.340921 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.348128 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.348169 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.348182 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.348198 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.348211 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.356763 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.373652 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.409128 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.428495 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.444682 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.454613 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.456350 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.456378 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.456396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.456407 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.459922 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.472556 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.482537 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.493121 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.559104 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.559138 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.559146 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.559165 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.559174 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.662190 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.662240 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.662249 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.662260 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.662270 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.766184 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.766254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.766270 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.766293 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.766312 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.869059 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.869108 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.869119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.869137 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.869149 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.948827 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 02:46:32.901670239 +0000 UTC Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.971128 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.971157 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.971166 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.971181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:15 crc kubenswrapper[4767]: I0129 15:01:15.971193 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:15Z","lastTransitionTime":"2026-01-29T15:01:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.018600 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.018680 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:16 crc kubenswrapper[4767]: E0129 15:01:16.018852 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:16 crc kubenswrapper[4767]: E0129 15:01:16.019058 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.074513 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.074605 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.074624 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.074652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.074674 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.178757 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.178854 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.178875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.178941 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.178965 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.242451 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.245881 4767 generic.go:334] "Generic (PLEG): container finished" podID="65812f82-4e2a-4f3d-bf4c-3aa608f3de8d" containerID="1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae" exitCode=0 Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.245941 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerDied","Data":"1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.267531 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.282019 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.282090 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.282111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.282139 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.282160 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.289756 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.302338 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.315288 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.336849 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.349341 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.360654 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.375374 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.384838 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.384918 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.384939 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.384964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.384982 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.391172 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.406267 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.420416 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.441107 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.454133 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.465841 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.488041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.488092 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.488105 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.488123 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.488135 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.491638 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:16Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.591836 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.591875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.591884 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.591931 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.591947 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.694822 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.694891 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.694946 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.694974 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.694996 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.798060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.798093 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.798102 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.798114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.798123 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.899922 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.899990 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.900000 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.900015 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.900027 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:16Z","lastTransitionTime":"2026-01-29T15:01:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:16 crc kubenswrapper[4767]: I0129 15:01:16.949856 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 03:42:39.763589188 +0000 UTC Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.003064 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.003133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.003156 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.003186 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.003208 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.018549 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:17 crc kubenswrapper[4767]: E0129 15:01:17.018690 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.105567 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.105615 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.105627 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.105642 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.105653 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.208159 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.208230 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.208254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.208284 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.208308 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.252152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerStarted","Data":"41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.269627 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.282867 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.299663 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.311341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.311377 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.311387 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.311401 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.311411 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.312971 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.326576 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.353199 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.368122 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.380175 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.399553 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.410366 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.413789 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.413833 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.413845 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.413864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.413875 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.422399 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.430653 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.439989 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.452129 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.462603 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:17Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.516552 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.516587 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.516620 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.516800 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.516815 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.620244 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.620350 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.620374 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.620403 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.620423 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.723480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.723534 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.723546 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.723567 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.723580 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.827751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.827822 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.827844 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.827875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.827925 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.930647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.930685 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.930697 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.930714 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.930725 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:17Z","lastTransitionTime":"2026-01-29T15:01:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:17 crc kubenswrapper[4767]: I0129 15:01:17.951060 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:29:14.442543229 +0000 UTC Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.017838 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.017934 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:18 crc kubenswrapper[4767]: E0129 15:01:18.018090 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:18 crc kubenswrapper[4767]: E0129 15:01:18.018366 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.032762 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.032828 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.032839 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.032855 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.032866 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.134964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.134998 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.135006 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.135020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.135029 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.237979 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.238012 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.238020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.238033 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.238042 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.257818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.258706 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.258757 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.264648 4767 generic.go:334] "Generic (PLEG): container finished" podID="65812f82-4e2a-4f3d-bf4c-3aa608f3de8d" containerID="41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd" exitCode=0 Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.264709 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerDied","Data":"41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.276098 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.291412 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.292129 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.297317 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.309218 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.321776 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.333911 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.340061 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.340095 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.340107 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.340123 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.340134 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.345333 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.355210 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.363418 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.373349 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.386312 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.397094 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.406513 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.417622 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.431086 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.443630 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.443714 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.443737 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.443769 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.443792 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.446183 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.469000 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.487866 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.502781 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.513003 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.526078 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.539209 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.546852 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.546908 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.546925 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.546941 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.546951 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.554559 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.570958 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.585532 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.598078 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.613811 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.633965 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.649624 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.649664 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.649673 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.649687 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.649697 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.654674 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.666463 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.678386 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:18Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.751522 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.751561 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.751572 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.751588 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.751598 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.855142 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.855531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.855565 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.857068 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.857091 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.951304 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 23:12:05.023297141 +0000 UTC Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.959512 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.959541 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.959553 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.959571 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:18 crc kubenswrapper[4767]: I0129 15:01:18.959582 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:18Z","lastTransitionTime":"2026-01-29T15:01:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.018189 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:19 crc kubenswrapper[4767]: E0129 15:01:19.018347 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.061936 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.061979 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.061989 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.062003 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.062016 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.164751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.164793 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.164806 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.164825 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.164842 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.266916 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.266957 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.266966 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.266979 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.266990 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.273553 4767 generic.go:334] "Generic (PLEG): container finished" podID="65812f82-4e2a-4f3d-bf4c-3aa608f3de8d" containerID="5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456" exitCode=0 Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.273664 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerDied","Data":"5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.273746 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.295754 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.320187 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.336081 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.353529 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.364810 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.370877 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.370949 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.370966 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.370986 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.371001 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.377357 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.400089 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.412767 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.427991 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.451459 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.462996 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.471264 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.472459 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.472551 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.472655 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.472764 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.472986 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.484355 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.502796 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.515238 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:19Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.619293 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.619341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.619366 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.619397 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.619420 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.722733 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.722767 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.722778 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.722795 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.722805 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.825629 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.825666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.825676 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.825690 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.825700 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.928719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.928747 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.928756 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.928768 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.928776 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:19Z","lastTransitionTime":"2026-01-29T15:01:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:19 crc kubenswrapper[4767]: I0129 15:01:19.951715 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 20:59:10.245500371 +0000 UTC Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.020570 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.020598 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:20 crc kubenswrapper[4767]: E0129 15:01:20.020730 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:20 crc kubenswrapper[4767]: E0129 15:01:20.021353 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.031781 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.031816 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.031823 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.031836 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.031845 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.139028 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.139069 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.139079 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.139098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.139112 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.241734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.242036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.242045 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.242057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.242066 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.279086 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" event={"ID":"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d","Type":"ContainerStarted","Data":"124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.279162 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.291498 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.303111 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.313643 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.323941 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.333930 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.344260 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.344306 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.344321 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.344342 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.344356 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.347570 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.363488 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.381292 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.394066 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.406394 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.420978 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.437607 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.446961 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.447010 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.447023 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.447037 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.447048 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.450970 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.459979 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.472665 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:20Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.550070 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.550114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.550126 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.550141 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.550150 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.652098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.652143 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.652151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.652165 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.652174 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.754302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.754337 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.754347 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.754363 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.754375 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.857499 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.857553 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.857564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.857580 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.857589 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.952496 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:40:09.232463396 +0000 UTC Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.960212 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.960261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.960276 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.960295 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:20 crc kubenswrapper[4767]: I0129 15:01:20.960309 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:20Z","lastTransitionTime":"2026-01-29T15:01:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.018066 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.018251 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.063447 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.063501 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.063514 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.063534 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.063546 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.167845 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.167958 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.167985 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.168017 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.168040 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.270316 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.270371 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.270386 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.270408 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.270420 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.285026 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/0.log" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.287827 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b" exitCode=1 Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.287875 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.288729 4767 scope.go:117] "RemoveContainer" containerID="c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.304115 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.319007 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.334427 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.344363 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.358439 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.369133 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.372563 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.372583 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.372591 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.372604 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.372613 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.378557 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.405361 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.436194 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.449753 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.460941 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.474220 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.474259 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.474268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.474284 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.474296 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.478619 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.495425 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.506638 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.518321 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.576709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.576746 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.576755 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.576769 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.576778 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.634412 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.634548 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.634606 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:01:37.634582864 +0000 UTC m=+53.444899913 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.634628 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.634660 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:37.634651876 +0000 UTC m=+53.444968915 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.680014 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.680068 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.680079 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.680094 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.680104 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.735205 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.735244 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.735265 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.735337 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.735381 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:37.735368337 +0000 UTC m=+53.545685376 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.735388 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.735417 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.735428 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.735477 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:37.73546067 +0000 UTC m=+53.545777709 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.735906 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.736011 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.736120 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.736254 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:37.736240262 +0000 UTC m=+53.546557321 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.782840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.782876 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.782887 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.782918 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.782930 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.825294 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.825322 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.825332 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.825344 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.825353 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.837462 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.842776 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.843024 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.843182 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.843325 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.843466 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.859025 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.865082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.865117 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.865127 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.865144 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.865156 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.879536 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.885515 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.885709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.885803 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.885887 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.886018 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.897418 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.901351 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.901398 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.901407 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.901421 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.901430 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.912568 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:21Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:21 crc kubenswrapper[4767]: E0129 15:01:21.912682 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.914306 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.914345 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.914354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.914370 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.914380 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:21Z","lastTransitionTime":"2026-01-29T15:01:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:21 crc kubenswrapper[4767]: I0129 15:01:21.952685 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 07:15:57.891969973 +0000 UTC Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.016934 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.016972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.016981 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.016994 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.017015 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.018060 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:22 crc kubenswrapper[4767]: E0129 15:01:22.018152 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.018178 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:22 crc kubenswrapper[4767]: E0129 15:01:22.018268 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.018680 4767 scope.go:117] "RemoveContainer" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.119545 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.119591 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.119602 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.119619 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.119631 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.222508 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.222551 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.222564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.222586 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.222599 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.293403 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/0.log" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.296009 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.296127 4767 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.310264 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.324223 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.325875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.325933 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.325945 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.325963 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.325974 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.359372 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.369771 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.381439 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.392715 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.405054 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.413995 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.425268 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.427675 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.427709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.427717 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.427732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.427742 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.436402 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.450451 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.469274 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.495391 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.506358 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.517411 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.530431 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.530483 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.530496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.530514 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.530526 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.633444 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.633481 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.633490 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.633505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.633514 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.735191 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.735231 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.735240 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.735264 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.735282 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.838511 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.838559 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.838572 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.838588 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.838600 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.905168 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr"] Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.905756 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.909225 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.909375 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.920364 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.935499 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.941622 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.941666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.941679 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.941695 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.941706 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:22Z","lastTransitionTime":"2026-01-29T15:01:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.953161 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 20:35:17.160574533 +0000 UTC Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.956511 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.975278 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:22 crc kubenswrapper[4767]: I0129 15:01:22.991932 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:22Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.003755 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.018383 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:23 crc kubenswrapper[4767]: E0129 15:01:23.018517 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.031500 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.044475 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.044526 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.044536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.044551 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.044561 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.046490 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.048463 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9c02de73-c364-474f-b2ff-86ed8e52530a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.048572 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9c02de73-c364-474f-b2ff-86ed8e52530a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.048611 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9c02de73-c364-474f-b2ff-86ed8e52530a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.048640 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7257m\" (UniqueName: \"kubernetes.io/projected/9c02de73-c364-474f-b2ff-86ed8e52530a-kube-api-access-7257m\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.064520 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.096478 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.112968 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.127881 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.136940 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.146655 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.146706 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.146714 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.146791 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.146802 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.149544 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9c02de73-c364-474f-b2ff-86ed8e52530a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.149652 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9c02de73-c364-474f-b2ff-86ed8e52530a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.149695 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9c02de73-c364-474f-b2ff-86ed8e52530a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.149658 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.149744 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7257m\" (UniqueName: \"kubernetes.io/projected/9c02de73-c364-474f-b2ff-86ed8e52530a-kube-api-access-7257m\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.150584 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9c02de73-c364-474f-b2ff-86ed8e52530a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.150604 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9c02de73-c364-474f-b2ff-86ed8e52530a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.155435 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9c02de73-c364-474f-b2ff-86ed8e52530a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.162219 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.166521 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7257m\" (UniqueName: \"kubernetes.io/projected/9c02de73-c364-474f-b2ff-86ed8e52530a-kube-api-access-7257m\") pod \"ovnkube-control-plane-749d76644c-qmhpr\" (UID: \"9c02de73-c364-474f-b2ff-86ed8e52530a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.174760 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.226444 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" Jan 29 15:01:23 crc kubenswrapper[4767]: W0129 15:01:23.238155 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c02de73_c364_474f_b2ff_86ed8e52530a.slice/crio-06776f977ab5bad8948de1144b61881a5bcb22d09949bedab5b31398b89b6208 WatchSource:0}: Error finding container 06776f977ab5bad8948de1144b61881a5bcb22d09949bedab5b31398b89b6208: Status 404 returned error can't find the container with id 06776f977ab5bad8948de1144b61881a5bcb22d09949bedab5b31398b89b6208 Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.248922 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.248972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.248984 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.249002 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.249016 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.301093 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/1.log" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.301596 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/0.log" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.309165 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4" exitCode=1 Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.309278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.309368 4767 scope.go:117] "RemoveContainer" containerID="c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.310114 4767 scope.go:117] "RemoveContainer" containerID="5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4" Jan 29 15:01:23 crc kubenswrapper[4767]: E0129 15:01:23.310274 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.311799 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" event={"ID":"9c02de73-c364-474f-b2ff-86ed8e52530a","Type":"ContainerStarted","Data":"06776f977ab5bad8948de1144b61881a5bcb22d09949bedab5b31398b89b6208"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.314070 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.317006 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.317786 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.322748 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.333165 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.342464 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.351684 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.351734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.351746 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.351767 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.351779 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.355056 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.366709 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.379888 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.398786 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.423500 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.435287 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.448859 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.454468 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.454496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.454505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.454520 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.454531 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.465257 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.477994 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.493721 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.510500 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.522152 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.536683 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.548954 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.556763 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.556787 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.556794 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.556808 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.556817 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.560611 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.571092 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.587559 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.600573 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.612844 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.643028 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.658781 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.658819 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.658831 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.658848 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.658861 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.660195 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.674760 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.696508 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.725010 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.736330 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.748457 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.758792 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.761141 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.761176 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.761185 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.761198 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.761208 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.772302 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.784011 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:23Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.864224 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.864524 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.864613 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.864698 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.864777 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.953655 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 14:34:24.154065115 +0000 UTC Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.967846 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.967939 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.967949 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.967969 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:23 crc kubenswrapper[4767]: I0129 15:01:23.967982 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:23Z","lastTransitionTime":"2026-01-29T15:01:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.018408 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.018452 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:24 crc kubenswrapper[4767]: E0129 15:01:24.018520 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:24 crc kubenswrapper[4767]: E0129 15:01:24.018713 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.070199 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.070273 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.070283 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.070303 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.070317 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.173280 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.173320 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.173332 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.173348 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.173361 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.275612 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.275649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.275662 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.275678 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.275689 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.321574 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" event={"ID":"9c02de73-c364-474f-b2ff-86ed8e52530a","Type":"ContainerStarted","Data":"f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.321623 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" event={"ID":"9c02de73-c364-474f-b2ff-86ed8e52530a","Type":"ContainerStarted","Data":"c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.324797 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/1.log" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.340840 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.366123 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.378771 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.378811 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.378821 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.378838 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.378850 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.386754 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.416064 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.426129 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.443115 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.456008 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.467699 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.480593 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.481179 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.481217 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.481225 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.481238 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.481247 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.491272 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.500188 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.509061 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.518551 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.526198 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.536367 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.549155 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.583791 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.583839 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.583857 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.583879 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.583919 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.691398 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.691435 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.691444 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.691457 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.691467 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.739891 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-lhmrn"] Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.740836 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:24 crc kubenswrapper[4767]: E0129 15:01:24.741007 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.765394 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.786034 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.794240 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.794278 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.794289 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.794307 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.794318 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.803411 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.840124 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.859524 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.867448 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkl5w\" (UniqueName: \"kubernetes.io/projected/03807b98-087e-4b91-b3ea-f2457587c580-kube-api-access-mkl5w\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.867488 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.871723 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.887282 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.897085 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.897131 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.897143 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.897156 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.897167 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.900221 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.915823 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.926694 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.936288 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.945064 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.954203 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.955248 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 22:08:23.164508142 +0000 UTC Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.968191 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkl5w\" (UniqueName: \"kubernetes.io/projected/03807b98-087e-4b91-b3ea-f2457587c580-kube-api-access-mkl5w\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.968236 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:24 crc kubenswrapper[4767]: E0129 15:01:24.968385 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:24 crc kubenswrapper[4767]: E0129 15:01:24.968440 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:25.468426143 +0000 UTC m=+41.278743182 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.975061 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.986708 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.993780 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkl5w\" (UniqueName: \"kubernetes.io/projected/03807b98-087e-4b91-b3ea-f2457587c580-kube-api-access-mkl5w\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.999056 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.999086 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.999098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.999114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:24 crc kubenswrapper[4767]: I0129 15:01:24.999125 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:24Z","lastTransitionTime":"2026-01-29T15:01:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:24.999972 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:24Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.021034 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:25 crc kubenswrapper[4767]: E0129 15:01:25.021179 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.021812 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.046063 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.057982 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.071343 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.082091 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.090887 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.099141 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.100988 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.101038 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.101050 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.101070 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.101083 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.121421 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.138178 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.149977 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.171793 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.185718 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.199424 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.202854 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.202961 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.202985 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.203013 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.203035 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.209991 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.226888 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.240664 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.252529 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.265150 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:25Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.305492 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.305529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.305539 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.305551 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.305559 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.409084 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.409152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.409177 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.409208 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.409231 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.474091 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:25 crc kubenswrapper[4767]: E0129 15:01:25.474244 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:25 crc kubenswrapper[4767]: E0129 15:01:25.474319 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:26.474296543 +0000 UTC m=+42.284613602 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.511817 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.511877 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.511929 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.511959 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.511981 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.614775 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.614840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.614863 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.614884 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.614920 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.718313 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.718380 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.718399 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.718420 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.718436 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.821628 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.821671 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.821683 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.821700 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.821712 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.925325 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.925367 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.925377 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.925393 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.925409 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:25Z","lastTransitionTime":"2026-01-29T15:01:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:25 crc kubenswrapper[4767]: I0129 15:01:25.955849 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 23:07:43.502640234 +0000 UTC Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.017952 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.017952 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:26 crc kubenswrapper[4767]: E0129 15:01:26.018136 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:26 crc kubenswrapper[4767]: E0129 15:01:26.018197 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.028998 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.029098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.029115 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.029140 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.029156 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.132027 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.132083 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.132096 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.132117 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.132132 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.234683 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.234721 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.234732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.234744 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.234753 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.337161 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.337222 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.337239 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.337264 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.337336 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.441280 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.441343 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.441366 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.441387 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.441401 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.484142 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:26 crc kubenswrapper[4767]: E0129 15:01:26.484350 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:26 crc kubenswrapper[4767]: E0129 15:01:26.484471 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:28.484439079 +0000 UTC m=+44.294756158 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.544831 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.544881 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.544913 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.544941 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.544955 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.648739 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.648828 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.648851 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.648884 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.648953 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.751574 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.751652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.751665 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.751681 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.751694 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.855371 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.855445 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.855464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.855489 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.855506 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.956840 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 18:53:38.6828403 +0000 UTC Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.958938 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.958996 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.959013 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.959041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:26 crc kubenswrapper[4767]: I0129 15:01:26.959059 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:26Z","lastTransitionTime":"2026-01-29T15:01:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.018549 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:27 crc kubenswrapper[4767]: E0129 15:01:27.018946 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.019036 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:27 crc kubenswrapper[4767]: E0129 15:01:27.019269 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.062479 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.062570 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.062594 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.062620 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.062641 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.165724 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.165786 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.165807 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.165833 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.165853 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.268483 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.268529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.268541 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.268556 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.268569 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.371469 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.371548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.371567 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.371593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.371611 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.474001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.474328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.474338 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.474354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.474368 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.576674 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.576734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.576748 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.576767 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.576783 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.679931 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.680000 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.680023 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.680052 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.680075 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.783099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.783160 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.783178 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.783239 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.783258 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.886250 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.886304 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.886323 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.886347 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.886362 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.957554 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 17:32:53.676773062 +0000 UTC Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.988606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.988666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.988687 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.988717 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:27 crc kubenswrapper[4767]: I0129 15:01:27.988739 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:27Z","lastTransitionTime":"2026-01-29T15:01:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.018338 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:28 crc kubenswrapper[4767]: E0129 15:01:28.018556 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.018338 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:28 crc kubenswrapper[4767]: E0129 15:01:28.018712 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.090890 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.091009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.091036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.091066 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.091090 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.193376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.193452 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.193475 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.193504 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.193526 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.296195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.296233 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.296264 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.296280 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.296292 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.399705 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.399784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.399809 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.399842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.399864 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.502786 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.502865 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.502889 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.502986 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.503010 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.507663 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:28 crc kubenswrapper[4767]: E0129 15:01:28.507840 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:28 crc kubenswrapper[4767]: E0129 15:01:28.507939 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:32.50791832 +0000 UTC m=+48.318235369 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.607169 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.607237 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.607254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.607279 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.607295 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.709878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.709987 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.710025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.710055 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.710076 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.813327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.813389 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.813409 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.813432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.813450 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.916436 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.916487 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.916501 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.916521 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.916536 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:28Z","lastTransitionTime":"2026-01-29T15:01:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:28 crc kubenswrapper[4767]: I0129 15:01:28.958185 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 10:44:10.08698105 +0000 UTC Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.017495 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:29 crc kubenswrapper[4767]: E0129 15:01:29.017645 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.017820 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:29 crc kubenswrapper[4767]: E0129 15:01:29.018054 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.019043 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.019095 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.019108 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.019126 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.019140 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.121657 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.121711 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.121728 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.121746 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.121759 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.224788 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.224874 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.224932 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.224972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.225014 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.328186 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.328255 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.328294 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.328327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.328349 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.431635 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.431702 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.431721 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.431743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.431759 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.534674 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.534723 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.534731 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.534745 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.534756 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.637521 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.637575 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.637588 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.637606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.637619 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.740713 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.740801 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.740842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.740889 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.740956 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.844941 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.845022 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.845046 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.845074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.845093 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.947784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.947825 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.947835 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.947850 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.947861 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:29Z","lastTransitionTime":"2026-01-29T15:01:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:29 crc kubenswrapper[4767]: I0129 15:01:29.958664 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 15:45:29.19050603 +0000 UTC Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.018338 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.018385 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:30 crc kubenswrapper[4767]: E0129 15:01:30.018479 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:30 crc kubenswrapper[4767]: E0129 15:01:30.018614 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.050497 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.050564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.050581 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.050606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.050622 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.153621 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.153685 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.153706 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.153729 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.153751 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.257308 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.257355 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.257366 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.257383 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.257394 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.359612 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.359645 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.359653 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.359668 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.359677 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.462743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.462804 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.462820 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.462850 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.462978 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.566419 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.566477 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.566494 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.566517 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.566533 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.670020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.670096 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.670114 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.670141 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.670160 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.772793 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.772881 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.772991 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.773029 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.773051 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.876652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.876723 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.876740 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.876764 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.876781 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.959017 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 02:16:59.268080596 +0000 UTC Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.980284 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.980340 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.980354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.980377 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:30 crc kubenswrapper[4767]: I0129 15:01:30.980394 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:30Z","lastTransitionTime":"2026-01-29T15:01:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.017936 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.018077 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:31 crc kubenswrapper[4767]: E0129 15:01:31.018180 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:31 crc kubenswrapper[4767]: E0129 15:01:31.018442 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.083068 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.083112 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.083123 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.083138 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.083148 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.186040 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.186099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.186116 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.186138 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.186155 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.288840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.288882 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.288907 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.288923 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.288933 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.391626 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.391662 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.391672 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.391692 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.391709 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.493471 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.493505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.493514 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.493527 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.493536 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.595844 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.595883 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.595891 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.595925 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.595936 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.699438 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.699530 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.699553 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.699583 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.699608 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.802252 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.802312 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.802329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.802354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.802371 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.905213 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.905252 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.905260 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.905276 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.905285 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:31Z","lastTransitionTime":"2026-01-29T15:01:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:31 crc kubenswrapper[4767]: I0129 15:01:31.959233 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 14:07:13.582680147 +0000 UTC Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.007327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.007359 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.007367 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.007379 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.007387 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.056883 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.057296 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.057390 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.057557 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.058345 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.058396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.058414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.058440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.058460 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.079647 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.084467 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.084522 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.084540 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.084559 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.084574 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.107004 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.112931 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.112982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.112999 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.113022 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.113039 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.132987 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.137996 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.138074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.138098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.138132 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.138156 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.160076 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.164387 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.164435 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.164448 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.164469 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.164486 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.179145 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.179300 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.181228 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.181265 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.181279 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.181300 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.181314 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.284332 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.284393 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.284411 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.284437 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.284458 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.386991 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.387052 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.387074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.387096 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.387115 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.490290 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.490395 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.490419 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.490465 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.490492 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.558343 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.558581 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:32 crc kubenswrapper[4767]: E0129 15:01:32.558687 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:40.558653727 +0000 UTC m=+56.368970816 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.586390 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.592498 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.592528 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.592539 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.592634 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.592687 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.608608 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.632655 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.648130 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.668068 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.685719 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.695667 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.696011 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.696088 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.696171 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.696264 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.701122 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.717009 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.731443 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.745642 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.759009 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.770229 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.779864 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.788969 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.798575 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.798610 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.798618 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.798634 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.798643 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.806415 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.817003 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.827726 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.843453 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:32Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.901471 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.901518 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.901531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.901547 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.901559 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:32Z","lastTransitionTime":"2026-01-29T15:01:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:32 crc kubenswrapper[4767]: I0129 15:01:32.959446 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 21:52:17.290758121 +0000 UTC Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.004929 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.005076 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.005105 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.005137 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.005160 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.018060 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.018125 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:33 crc kubenswrapper[4767]: E0129 15:01:33.018289 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:33 crc kubenswrapper[4767]: E0129 15:01:33.018429 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.107578 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.107649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.107667 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.107688 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.107702 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.211072 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.211136 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.211148 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.211168 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.211185 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.302232 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.317822 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.317862 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.317873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.317905 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.317920 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.322150 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.330370 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.358784 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.383223 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.397797 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.412858 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.420812 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.420868 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.420880 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.420916 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.420935 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.432005 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.444353 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.458225 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.473415 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.484278 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.494410 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.506880 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.517957 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.522807 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.522841 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.522852 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.522869 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.522881 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.526624 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.536256 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.548042 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.562906 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:33Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.624873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.624919 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.624927 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.624939 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.624948 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.727710 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.727772 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.727796 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.727819 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.727840 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.829607 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.829649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.829660 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.829676 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.829688 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.931316 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.931352 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.931364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.931378 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.931388 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:33Z","lastTransitionTime":"2026-01-29T15:01:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:33 crc kubenswrapper[4767]: I0129 15:01:33.959687 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 16:49:06.905123095 +0000 UTC Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.017713 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.017722 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:34 crc kubenswrapper[4767]: E0129 15:01:34.017927 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:34 crc kubenswrapper[4767]: E0129 15:01:34.017964 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.033495 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.033532 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.033540 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.033557 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.033566 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.137391 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.137451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.137472 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.137500 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.137524 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.240289 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.240348 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.240364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.240388 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.240402 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.342678 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.342726 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.342738 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.342755 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.342767 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.445110 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.445169 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.445184 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.445232 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.445245 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.547128 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.547172 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.547181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.547193 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.547202 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.650332 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.650380 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.650394 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.650409 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.650419 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.752720 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.752765 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.752778 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.752792 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.752800 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.854788 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.854868 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.854945 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.854985 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.855009 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.957178 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.957243 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.957263 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.957289 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.957306 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:34Z","lastTransitionTime":"2026-01-29T15:01:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:34 crc kubenswrapper[4767]: I0129 15:01:34.960369 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 18:41:48.028988686 +0000 UTC Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.018467 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:35 crc kubenswrapper[4767]: E0129 15:01:35.018607 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.018715 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:35 crc kubenswrapper[4767]: E0129 15:01:35.018871 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.031343 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.042300 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.054200 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.059878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.059926 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.059937 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.059954 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.059965 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.067480 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.077825 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.088486 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.101537 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.118276 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.131831 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.144233 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.161226 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d030f2a6dff994162cc8d38cd81320c51c629decc2c5d96f6da565a41bbf2b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"message\\\":\\\"go/informers/factory.go:160\\\\nI0129 15:01:20.810639 6017 reflector.go:311] Stopping reflector *v1.UserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.810821 6017 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811237 6017 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811262 6017 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0129 15:01:20.811299 6017 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0129 15:01:20.811762 6017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0129 15:01:20.811797 6017 factory.go:656] Stopping watch factory\\\\nI0129 15:01:20.811808 6017 ovnkube.go:599] Stopped ovnkube\\\\nI0129 15:01:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.162837 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.162879 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.162913 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.162932 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.162950 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.177788 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.188012 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.202255 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.216964 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.229450 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.241341 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.255433 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:35Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.265660 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.265691 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.265699 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.265712 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.265722 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.367488 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.367549 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.367568 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.367592 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.367608 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.469993 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.470043 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.470056 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.470077 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.470092 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.573992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.574072 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.574099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.574133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.574158 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.677324 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.677437 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.677501 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.677523 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.677560 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.780484 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.780526 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.780534 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.780548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.780557 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.883019 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.883065 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.883080 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.883103 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.883120 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.961059 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 06:48:27.380514505 +0000 UTC Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.984953 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.985014 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.985026 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.985041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:35 crc kubenswrapper[4767]: I0129 15:01:35.985051 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:35Z","lastTransitionTime":"2026-01-29T15:01:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.008490 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.009680 4767 scope.go:117] "RemoveContainer" containerID="5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.018467 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:36 crc kubenswrapper[4767]: E0129 15:01:36.018561 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.018718 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:36 crc kubenswrapper[4767]: E0129 15:01:36.019221 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.025992 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.038719 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.056040 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.067044 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.077694 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.087576 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.087646 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.087670 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.087698 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.087724 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.088686 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.103976 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.117496 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.132463 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.150551 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.173110 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.190758 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.190793 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.190804 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.190818 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.190827 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.200106 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.213232 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.228755 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.243780 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.256655 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.273099 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.288457 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.293827 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.293863 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.293875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.293905 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.293918 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.370549 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/1.log" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.373356 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.374434 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.395031 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.396188 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.396240 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.396253 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.396274 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.396288 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.413676 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.434965 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.456456 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.469756 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.483009 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.494806 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.498492 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.498635 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.498705 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.498769 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.498833 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.508652 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.522523 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.537085 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.550943 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.562057 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.576105 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.590464 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.602084 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.602163 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.602180 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.602203 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.602226 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.613450 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.636044 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.656401 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.672850 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:36Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.704909 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.704942 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.704952 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.704965 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.704974 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.807344 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.807376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.807384 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.807396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.807404 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.910549 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.910589 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.910602 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.910618 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.910628 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:36Z","lastTransitionTime":"2026-01-29T15:01:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:36 crc kubenswrapper[4767]: I0129 15:01:36.961297 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 12:54:13.028382116 +0000 UTC Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.013310 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.013557 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.013574 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.013597 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.013616 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.020133 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.020287 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.020394 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.020594 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.116627 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.116666 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.116679 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.116698 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.116708 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.219422 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.219466 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.219479 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.219498 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.219510 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.322405 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.322443 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.322452 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.322465 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.322474 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.378541 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/2.log" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.379421 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/1.log" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.381854 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089" exitCode=1 Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.381936 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.382055 4767 scope.go:117] "RemoveContainer" containerID="5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.382540 4767 scope.go:117] "RemoveContainer" containerID="3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089" Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.382692 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.396554 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.410532 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.423851 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.424841 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.424864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.424873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.424886 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.424909 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.432511 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.442817 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.455314 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.465711 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.474136 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.481953 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.493292 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.505304 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.517296 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.526989 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.527032 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.527045 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.527060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.527072 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.528238 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.542146 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.559988 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.571072 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.583139 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.601996 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5470fc0470a6c32448fa04bbc549f3fa1b8709151e094d03e2c07812605864d4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"message\\\":\\\"api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"de88cb48-af91-44f8-b3c0-73dcf8201ba5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF0129 15:01:21.997635 6191 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:37Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.629686 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.629717 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.629724 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.629738 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.629746 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.705311 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.705454 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.705565 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.705629 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:02:09.705611158 +0000 UTC m=+85.515928197 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.705714 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:02:09.70570081 +0000 UTC m=+85.516017849 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.732183 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.732220 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.732230 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.732243 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.732253 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.806860 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807100 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807215 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807230 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.807185 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807284 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:02:09.807269205 +0000 UTC m=+85.617586244 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.807337 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807402 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807428 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:02:09.807420629 +0000 UTC m=+85.617737668 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807560 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807637 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807725 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:37 crc kubenswrapper[4767]: E0129 15:01:37.807836 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:02:09.80782386 +0000 UTC m=+85.618140899 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.834180 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.834215 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.834223 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.834237 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.834248 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.936561 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.936807 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.936880 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.936976 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.937055 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:37Z","lastTransitionTime":"2026-01-29T15:01:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:37 crc kubenswrapper[4767]: I0129 15:01:37.961753 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 19:21:21.20094352 +0000 UTC Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.017445 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:38 crc kubenswrapper[4767]: E0129 15:01:38.017544 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.017445 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:38 crc kubenswrapper[4767]: E0129 15:01:38.017866 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.039120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.039150 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.039159 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.039174 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.039185 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.141701 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.141727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.141735 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.141747 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.141757 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.244813 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.245127 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.245187 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.245246 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.245314 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.347241 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.347296 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.347312 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.347327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.347337 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.387810 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/2.log" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.391204 4767 scope.go:117] "RemoveContainer" containerID="3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089" Jan 29 15:01:38 crc kubenswrapper[4767]: E0129 15:01:38.391416 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.404064 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.415444 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.427376 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.439677 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.449511 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.449654 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.449701 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.449712 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.449727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.449735 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.460862 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.471981 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.489866 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.502650 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.513724 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.529982 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.550928 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.553274 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.553302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.553313 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.553327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.553338 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.563637 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.573661 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.584877 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.598617 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.608523 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.618516 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:38Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.655291 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.655331 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.655341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.655355 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.655366 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.757103 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.757130 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.757139 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.757151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.757159 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.858851 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.858882 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.858891 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.858923 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.858934 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.961942 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 19:18:42.512276492 +0000 UTC Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.961992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.962041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.962057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.962077 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:38 crc kubenswrapper[4767]: I0129 15:01:38.962090 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:38Z","lastTransitionTime":"2026-01-29T15:01:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.018633 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:39 crc kubenswrapper[4767]: E0129 15:01:39.018765 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.019200 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:39 crc kubenswrapper[4767]: E0129 15:01:39.019284 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.064757 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.065009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.065099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.065198 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.065267 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.167606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.167648 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.167659 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.167673 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.167684 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.270947 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.270996 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.271013 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.271035 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.271054 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.373811 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.374204 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.374428 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.374587 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.374729 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.477305 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.477348 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.477360 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.477377 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.477388 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.580690 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.580732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.580743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.580760 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.580771 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.683862 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.683956 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.684011 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.684039 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.684057 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.787590 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.787656 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.787674 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.787703 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.787722 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.891649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.891721 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.891748 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.891776 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.891796 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.963219 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 04:47:15.826948832 +0000 UTC Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.994219 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.994279 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.994296 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.994320 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:39 crc kubenswrapper[4767]: I0129 15:01:39.994337 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:39Z","lastTransitionTime":"2026-01-29T15:01:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.018198 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.018249 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:40 crc kubenswrapper[4767]: E0129 15:01:40.018405 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:40 crc kubenswrapper[4767]: E0129 15:01:40.018580 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.108003 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.108725 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.109195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.109433 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.109661 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.213107 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.213642 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.213854 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.214111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.214327 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.317958 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.318031 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.318056 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.318086 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.318106 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.420822 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.420864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.420872 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.420886 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.420930 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.524002 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.524100 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.524120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.524145 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.524164 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.627240 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.627295 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.627309 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.627331 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.627347 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.636239 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:40 crc kubenswrapper[4767]: E0129 15:01:40.636454 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:40 crc kubenswrapper[4767]: E0129 15:01:40.636585 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:01:56.636564481 +0000 UTC m=+72.446881530 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.730315 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.730370 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.730388 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.730411 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.730429 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.833485 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.833577 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.833602 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.833629 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.833650 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.937334 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.937373 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.937383 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.937400 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.937412 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:40Z","lastTransitionTime":"2026-01-29T15:01:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:40 crc kubenswrapper[4767]: I0129 15:01:40.963692 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 16:28:38.106213245 +0000 UTC Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.017638 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.017755 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:41 crc kubenswrapper[4767]: E0129 15:01:41.017818 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:41 crc kubenswrapper[4767]: E0129 15:01:41.017962 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.040341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.040417 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.040442 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.040476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.040501 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.144332 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.144386 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.144401 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.144421 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.144436 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.246935 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.246997 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.247014 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.247035 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.247048 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.349393 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.349430 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.349439 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.349451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.349461 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.452152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.452205 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.452220 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.452243 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.452258 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.554885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.554982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.555000 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.555018 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.555031 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.657299 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.657367 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.657395 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.657436 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.657462 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.760517 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.760560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.760571 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.760587 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.760597 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.863480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.863514 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.863523 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.863539 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.863550 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.964015 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 02:29:32.057949607 +0000 UTC Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.966392 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.966438 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.966456 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.966486 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:41 crc kubenswrapper[4767]: I0129 15:01:41.966512 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:41Z","lastTransitionTime":"2026-01-29T15:01:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.018124 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.018281 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.018485 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.018542 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.073645 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.073727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.073763 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.073797 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.073821 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.178320 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.178373 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.178384 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.178402 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.178414 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.282130 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.282222 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.282235 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.282252 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.282265 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.289715 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.289743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.289753 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.289765 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.289775 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.309792 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:42Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.313944 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.313991 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.314007 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.314028 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.314043 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.333218 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:42Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.337693 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.337728 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.337741 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.337756 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.337768 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.354034 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:42Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.357567 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.357601 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.357613 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.357626 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.357637 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.372090 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:42Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.375750 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.375795 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.375808 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.375823 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.375837 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.387653 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:42Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:42 crc kubenswrapper[4767]: E0129 15:01:42.387765 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.389057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.389094 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.389105 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.389122 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.389134 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.491965 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.492020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.492036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.492057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.492072 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.595878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.595973 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.595995 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.596027 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.596049 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.698953 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.699048 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.699068 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.699093 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.699110 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.801594 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.801659 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.801670 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.801690 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.801702 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.904981 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.905031 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.905046 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.905064 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.905078 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:42Z","lastTransitionTime":"2026-01-29T15:01:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:42 crc kubenswrapper[4767]: I0129 15:01:42.964335 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 04:02:00.734496145 +0000 UTC Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.013328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.013388 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.013399 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.013414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.013428 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.017915 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:43 crc kubenswrapper[4767]: E0129 15:01:43.018026 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.018794 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:43 crc kubenswrapper[4767]: E0129 15:01:43.018933 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.115667 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.115710 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.115720 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.115734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.115747 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.219255 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.219313 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.219331 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.219351 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.219365 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.321689 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.321777 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.321801 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.321830 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.321853 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.424830 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.424955 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.424993 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.425032 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.425059 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.528302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.528392 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.528410 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.528436 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.528458 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.631510 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.631587 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.631605 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.631632 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.631652 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.735647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.735734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.735751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.735766 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.735777 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.839053 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.839117 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.839133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.839154 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.839169 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.941661 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.941706 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.941720 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.941742 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.941755 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:43Z","lastTransitionTime":"2026-01-29T15:01:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:43 crc kubenswrapper[4767]: I0129 15:01:43.965204 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 00:24:03.048236976 +0000 UTC Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.018198 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.018197 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:44 crc kubenswrapper[4767]: E0129 15:01:44.018474 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:44 crc kubenswrapper[4767]: E0129 15:01:44.018603 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.044361 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.044447 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.044481 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.044510 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.044532 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.147982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.148070 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.148094 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.148122 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.148143 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.252178 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.252268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.252305 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.252335 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.252356 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.355109 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.355147 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.355157 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.355172 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.355183 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.458185 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.458232 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.458245 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.458261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.458274 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.561246 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.561329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.561353 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.561383 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.561404 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.665213 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.665261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.665286 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.665332 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.665355 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.768589 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.768635 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.768646 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.768661 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.768673 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.871648 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.871704 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.871719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.871739 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.871754 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.965792 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 10:58:07.553005647 +0000 UTC Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.973626 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.973672 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.973682 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.973698 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:44 crc kubenswrapper[4767]: I0129 15:01:44.973712 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:44Z","lastTransitionTime":"2026-01-29T15:01:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.018271 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.018397 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:45 crc kubenswrapper[4767]: E0129 15:01:45.018473 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:45 crc kubenswrapper[4767]: E0129 15:01:45.018537 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.032430 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.045991 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.059005 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.072653 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.076586 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.076609 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.076617 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.076631 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.076640 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.084507 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.097203 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.109190 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.118611 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.138496 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.150415 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.163999 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.178147 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.178184 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.178196 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.178212 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.178225 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.192316 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.207135 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.222937 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.233389 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.245583 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.260602 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.272597 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:45Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.280631 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.280671 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.280681 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.280697 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.280711 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.383808 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.383870 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.383889 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.383940 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.383959 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.487297 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.487667 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.487680 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.487701 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.487714 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.590548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.590631 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.590650 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.590673 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.590691 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.694216 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.694258 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.694267 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.694282 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.694294 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.796589 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.796625 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.796634 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.796647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.796655 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.899865 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.899918 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.899928 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.899942 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.899952 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:45Z","lastTransitionTime":"2026-01-29T15:01:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:45 crc kubenswrapper[4767]: I0129 15:01:45.966981 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 15:28:12.333686084 +0000 UTC Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.002944 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.002996 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.003010 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.003029 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.003044 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.018527 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.018551 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:46 crc kubenswrapper[4767]: E0129 15:01:46.018653 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:46 crc kubenswrapper[4767]: E0129 15:01:46.018738 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.105247 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.105287 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.105300 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.105317 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.105329 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.207129 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.207181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.207195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.207214 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.207230 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.310458 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.310513 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.310529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.310549 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.310563 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.413402 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.413442 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.413451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.413463 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.413470 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.516153 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.516207 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.516225 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.516249 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.516264 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.619707 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.619752 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.619774 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.619793 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.619806 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.723420 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.723485 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.723496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.723516 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.723529 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.828969 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.829038 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.829057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.829084 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.829107 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.932957 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.933009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.933025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.933050 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.933066 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:46Z","lastTransitionTime":"2026-01-29T15:01:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:46 crc kubenswrapper[4767]: I0129 15:01:46.967884 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 12:08:33.612249095 +0000 UTC Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.018248 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:47 crc kubenswrapper[4767]: E0129 15:01:47.018486 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.018567 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:47 crc kubenswrapper[4767]: E0129 15:01:47.018773 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.036329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.036505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.036531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.036555 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.036568 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.140057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.140125 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.140145 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.140172 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.140190 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.243990 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.244043 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.244056 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.244074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.244086 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.347673 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.347716 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.347730 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.347748 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.347763 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.450434 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.450480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.450490 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.450510 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.450521 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.552661 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.552707 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.552719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.552736 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.552748 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.655420 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.655464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.655476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.655496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.655508 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.758317 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.758356 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.758364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.758380 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.758404 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.862161 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.862218 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.862234 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.862253 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.862267 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.965279 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.965340 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.965354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.965372 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.965384 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:47Z","lastTransitionTime":"2026-01-29T15:01:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:47 crc kubenswrapper[4767]: I0129 15:01:47.968453 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 13:55:47.676181011 +0000 UTC Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.017965 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.018019 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:48 crc kubenswrapper[4767]: E0129 15:01:48.018183 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:48 crc kubenswrapper[4767]: E0129 15:01:48.018275 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.068170 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.068226 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.068241 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.068263 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.068278 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.171472 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.171504 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.171512 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.171525 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.171535 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.274355 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.274394 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.274403 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.274415 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.274425 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.376415 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.376444 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.376453 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.376471 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.376488 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.478308 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.478338 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.478345 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.478357 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.478366 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.580395 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.580436 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.580444 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.580458 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.580468 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.682649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.682697 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.682709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.682723 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.682732 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.785484 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.785529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.785542 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.785560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.785573 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.887801 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.887850 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.887863 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.887881 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.887925 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.969475 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 08:10:03.265370914 +0000 UTC Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.990839 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.990933 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.990946 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.990964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:48 crc kubenswrapper[4767]: I0129 15:01:48.990975 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:48Z","lastTransitionTime":"2026-01-29T15:01:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.018644 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.018708 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:49 crc kubenswrapper[4767]: E0129 15:01:49.018834 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:49 crc kubenswrapper[4767]: E0129 15:01:49.018985 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.094640 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.094695 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.094710 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.094738 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.094753 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.197413 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.197464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.197478 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.197499 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.197514 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.300544 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.300607 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.300620 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.300637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.300649 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.402724 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.402750 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.402759 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.402774 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.402784 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.505219 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.505257 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.505268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.505283 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.505294 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.607687 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.607721 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.607730 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.607743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.607753 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.710466 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.710508 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.710517 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.710531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.710541 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.813296 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.813371 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.813383 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.813401 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.813413 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.916413 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.916475 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.916499 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.916530 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.916554 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:49Z","lastTransitionTime":"2026-01-29T15:01:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:49 crc kubenswrapper[4767]: I0129 15:01:49.969865 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 16:44:37.54094458 +0000 UTC Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.017523 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.017659 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:50 crc kubenswrapper[4767]: E0129 15:01:50.017716 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:50 crc kubenswrapper[4767]: E0129 15:01:50.019243 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.019470 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.019504 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.019516 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.019531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.019542 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.121227 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.121272 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.121288 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.121310 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.121331 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.225972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.226089 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.226104 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.226121 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.226512 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.329233 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.329275 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.329286 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.329301 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.329313 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.431539 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.431575 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.431584 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.431598 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.431608 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.533973 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.534019 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.534029 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.534046 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.534057 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.636211 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.636239 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.636247 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.636259 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.636269 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.739339 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.739383 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.739397 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.739415 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.739430 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.841644 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.841674 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.841685 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.841700 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.841713 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.944149 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.944184 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.944192 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.944207 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.944215 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:50Z","lastTransitionTime":"2026-01-29T15:01:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:50 crc kubenswrapper[4767]: I0129 15:01:50.970345 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 13:21:02.519167381 +0000 UTC Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.018135 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:51 crc kubenswrapper[4767]: E0129 15:01:51.018282 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.018395 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:51 crc kubenswrapper[4767]: E0129 15:01:51.018871 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.019022 4767 scope.go:117] "RemoveContainer" containerID="3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089" Jan 29 15:01:51 crc kubenswrapper[4767]: E0129 15:01:51.019189 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.046943 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.046982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.046991 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.047005 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.047017 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.149743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.149786 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.149797 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.149811 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.149819 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.252541 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.252583 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.252593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.252609 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.252618 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.354848 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.354885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.354916 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.354933 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.354946 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.457672 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.457720 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.457737 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.457753 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.457766 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.560073 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.560112 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.560120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.560134 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.560146 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.662328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.662370 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.662378 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.662391 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.662401 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.765846 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.765926 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.765940 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.765957 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.765969 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.867969 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.868016 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.868028 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.868044 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.868059 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.970599 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 21:56:01.073846748 +0000 UTC Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.970647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.970705 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.970723 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.970743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:51 crc kubenswrapper[4767]: I0129 15:01:51.970755 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:51Z","lastTransitionTime":"2026-01-29T15:01:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.018409 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.018542 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.018577 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.018733 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.073760 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.073790 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.073798 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.073810 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.073819 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.176536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.176591 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.176599 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.176614 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.176625 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.279411 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.279454 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.279463 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.279476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.279484 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.381934 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.381975 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.381986 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.382001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.382009 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.587236 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.587293 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.587306 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.587323 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.587338 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.588569 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.588608 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.588622 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.588639 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.588650 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.604111 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:52Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.607184 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.607214 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.607223 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.607238 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.607255 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.617828 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:52Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.620647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.620691 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.620702 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.620719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.620730 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.631007 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:52Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.633744 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.633776 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.633787 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.633805 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.633818 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.644270 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:52Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.647479 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.647512 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.647524 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.647540 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.647552 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.658491 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:52Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:52 crc kubenswrapper[4767]: E0129 15:01:52.658854 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.689026 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.689058 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.689067 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.689080 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.689091 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.791907 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.791949 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.791959 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.791973 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.791982 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.894382 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.894419 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.894430 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.894445 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.894455 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.971110 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 14:50:54.242480813 +0000 UTC Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.996354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.996399 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.996407 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.996421 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:52 crc kubenswrapper[4767]: I0129 15:01:52.996431 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:52Z","lastTransitionTime":"2026-01-29T15:01:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.017654 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.017766 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:53 crc kubenswrapper[4767]: E0129 15:01:53.017868 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:53 crc kubenswrapper[4767]: E0129 15:01:53.018465 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.098945 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.098992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.099001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.099014 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.099023 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.201464 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.201515 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.201529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.201547 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.201559 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.303604 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.303641 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.303652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.303668 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.303679 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.406122 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.406167 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.406178 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.406197 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.406210 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.508695 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.508730 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.508740 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.508754 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.508764 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.610550 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.610578 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.610589 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.610632 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.610644 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.713149 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.713187 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.713195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.713209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.713220 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.815233 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.815288 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.815305 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.815329 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.815345 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.917803 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.917859 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.917875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.917929 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.917945 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:53Z","lastTransitionTime":"2026-01-29T15:01:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:53 crc kubenswrapper[4767]: I0129 15:01:53.972301 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 06:57:36.255206534 +0000 UTC Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.018178 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.018178 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:54 crc kubenswrapper[4767]: E0129 15:01:54.018313 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:54 crc kubenswrapper[4767]: E0129 15:01:54.018395 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.019596 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.019626 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.019637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.019650 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.019660 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.121875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.121952 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.121965 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.121982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.121995 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.224430 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.224467 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.224477 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.224492 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.224503 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.326889 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.326972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.326984 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.327001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.327012 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.429389 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.429474 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.429489 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.429505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.429516 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.531557 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.531593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.531601 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.531614 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.531626 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.633977 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.634020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.634032 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.634074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.634086 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.736982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.737028 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.737036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.737051 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.737061 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.839210 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.839241 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.839249 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.839261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.839271 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.942212 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.942275 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.942298 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.942324 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.942343 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:54Z","lastTransitionTime":"2026-01-29T15:01:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:54 crc kubenswrapper[4767]: I0129 15:01:54.973484 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 02:21:28.718485964 +0000 UTC Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.018231 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:55 crc kubenswrapper[4767]: E0129 15:01:55.018623 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.018252 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:55 crc kubenswrapper[4767]: E0129 15:01:55.019285 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.034036 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.044483 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.044520 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.044530 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.044545 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.044556 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.056177 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.065870 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.076857 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.090291 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.100419 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.112054 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.123495 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.134204 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.145090 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.146658 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.146705 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.146717 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.146735 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.146748 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.155352 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.165845 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.175349 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.183380 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.199078 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.212527 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.223348 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.243253 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:55Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.248668 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.248724 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.248735 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.248748 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.248757 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.350600 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.350645 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.350655 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.350672 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.350683 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.452447 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.452494 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.452503 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.452517 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.452526 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.555315 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.555365 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.555381 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.555402 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.555417 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.657474 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.657524 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.657538 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.657556 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.657568 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.760119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.760169 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.760180 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.760197 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.760208 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.861980 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.862057 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.862074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.862096 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.862113 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.965065 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.965119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.965132 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.965150 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.965162 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:55Z","lastTransitionTime":"2026-01-29T15:01:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:55 crc kubenswrapper[4767]: I0129 15:01:55.974230 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 02:25:53.189374088 +0000 UTC Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.018363 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.018385 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:56 crc kubenswrapper[4767]: E0129 15:01:56.018496 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:56 crc kubenswrapper[4767]: E0129 15:01:56.018618 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.068431 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.068508 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.068521 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.068564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.068579 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.171810 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.171860 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.171873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.171889 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.171916 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.274763 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.274810 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.274821 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.274837 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.274849 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.377511 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.377575 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.377593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.377616 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.377633 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.480453 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.480554 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.480577 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.480600 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.480617 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.583244 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.583287 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.583369 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.583388 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.583399 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.685849 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.686023 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.686053 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.686082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.686101 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.723204 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:56 crc kubenswrapper[4767]: E0129 15:01:56.723367 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:56 crc kubenswrapper[4767]: E0129 15:01:56.723479 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:02:28.72345188 +0000 UTC m=+104.533768959 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.788360 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.788404 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.788420 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.788440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.788456 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.890433 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.890470 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.890480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.890492 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.890501 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.974441 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 13:27:29.150536019 +0000 UTC Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.993290 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.993367 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.993380 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.993396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:56 crc kubenswrapper[4767]: I0129 15:01:56.993431 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:56Z","lastTransitionTime":"2026-01-29T15:01:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.017556 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:57 crc kubenswrapper[4767]: E0129 15:01:57.017765 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.017839 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:57 crc kubenswrapper[4767]: E0129 15:01:57.018031 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.096577 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.096624 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.096635 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.096651 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.096662 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.198743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.198787 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.198798 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.198812 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.198822 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.301430 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.301461 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.301469 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.301482 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.301489 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.403378 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.403420 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.403431 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.403450 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.403463 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.505587 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.505637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.505648 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.505664 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.505676 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.607817 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.607875 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.607914 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.607939 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.607956 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.710379 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.710484 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.710506 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.710527 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.710578 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.812755 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.812816 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.812835 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.812858 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.812875 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.915606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.915635 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.915644 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.915657 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.915666 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:57Z","lastTransitionTime":"2026-01-29T15:01:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:57 crc kubenswrapper[4767]: I0129 15:01:57.975363 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 21:37:03.886257151 +0000 UTC Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.017443 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.017480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.017489 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.017503 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.017506 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.017544 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:01:58 crc kubenswrapper[4767]: E0129 15:01:58.017620 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.017513 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: E0129 15:01:58.017801 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.120218 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.120258 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.120271 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.120285 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.120296 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.222770 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.222815 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.222827 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.222844 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.222856 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.324922 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.324974 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.324985 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.325000 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.325009 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.427554 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.427585 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.427593 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.427608 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.427618 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.530602 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.530652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.530662 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.530677 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.530686 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.605118 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/0.log" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.605162 4767 generic.go:334] "Generic (PLEG): container finished" podID="4615231a-e157-430e-9ecc-97c3fd7701bb" containerID="4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d" exitCode=1 Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.605190 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerDied","Data":"4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.605539 4767 scope.go:117] "RemoveContainer" containerID="4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.622340 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.637533 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.637592 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.637610 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.637633 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.637652 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.646394 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.658774 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.674160 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"2026-01-29T15:01:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f\\\\n2026-01-29T15:01:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f to /host/opt/cni/bin/\\\\n2026-01-29T15:01:13Z [verbose] multus-daemon started\\\\n2026-01-29T15:01:13Z [verbose] Readiness Indicator file check\\\\n2026-01-29T15:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.691934 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.704865 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.722644 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.738868 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.741470 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.742111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.742311 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.742651 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.742966 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.751803 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.764683 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.777623 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.788101 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.797595 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.807657 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.825555 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.841018 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.845660 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.845686 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.845694 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.845707 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.845717 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.852982 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.873287 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:58Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.947586 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.947631 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.947640 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.947656 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.947666 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:58Z","lastTransitionTime":"2026-01-29T15:01:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:58 crc kubenswrapper[4767]: I0129 15:01:58.975954 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 02:07:48.456448136 +0000 UTC Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.017738 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.017876 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:01:59 crc kubenswrapper[4767]: E0129 15:01:59.017942 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:01:59 crc kubenswrapper[4767]: E0129 15:01:59.018128 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.049557 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.049599 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.049609 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.049623 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.049632 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.151912 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.151956 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.151965 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.151978 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.151986 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.254440 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.254476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.254485 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.254500 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.254514 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.357172 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.357213 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.357222 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.357236 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.357246 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.460106 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.460173 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.460186 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.460203 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.460214 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.563328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.563377 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.563428 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.563451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.563509 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.610231 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/0.log" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.610307 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerStarted","Data":"37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.625291 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.642614 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.656882 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.666480 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.666539 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.666548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.666562 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.666570 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.670824 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.683861 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.697970 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.708601 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.720939 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.741770 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.752255 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.763521 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.772693 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.772731 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.772743 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.772759 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.772771 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.782627 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.799964 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.811395 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.820202 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.830547 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"2026-01-29T15:01:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f\\\\n2026-01-29T15:01:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f to /host/opt/cni/bin/\\\\n2026-01-29T15:01:13Z [verbose] multus-daemon started\\\\n2026-01-29T15:01:13Z [verbose] Readiness Indicator file check\\\\n2026-01-29T15:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.842666 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.854097 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:01:59Z is after 2025-08-24T17:21:41Z" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.875142 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.875210 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.875221 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.875247 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.875262 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.976168 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 18:07:57.503427766 +0000 UTC Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.977572 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.977627 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.977638 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.977652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:01:59 crc kubenswrapper[4767]: I0129 15:01:59.977661 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:01:59Z","lastTransitionTime":"2026-01-29T15:01:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.018821 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.018821 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:00 crc kubenswrapper[4767]: E0129 15:02:00.019013 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:00 crc kubenswrapper[4767]: E0129 15:02:00.019086 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.080047 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.080151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.080177 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.080208 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.080230 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.182876 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.182938 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.182949 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.182967 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.182978 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.285614 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.285675 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.285691 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.285709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.285722 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.387755 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.387806 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.387820 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.387837 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.387848 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.490121 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.490150 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.490159 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.490171 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.490179 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.592432 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.592465 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.592472 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.592485 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.592493 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.694962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.694998 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.695009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.695025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.695036 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.797378 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.797455 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.797478 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.797508 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.797529 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.899699 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.899732 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.899747 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.899764 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.899774 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:00Z","lastTransitionTime":"2026-01-29T15:02:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:00 crc kubenswrapper[4767]: I0129 15:02:00.976963 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 00:49:07.959015551 +0000 UTC Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.003156 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.003257 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.003292 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.003316 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.003335 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.019160 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.019355 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:01 crc kubenswrapper[4767]: E0129 15:02:01.019498 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:01 crc kubenswrapper[4767]: E0129 15:02:01.019777 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.032843 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.105791 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.105830 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.105840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.105853 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.105863 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.209180 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.209337 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.209364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.209389 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.209447 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.311607 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.311658 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.311667 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.311682 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.311698 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.413769 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.413831 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.413854 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.413878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.413924 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.516670 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.516726 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.516738 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.516757 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.516770 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.618759 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.618811 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.618823 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.618842 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.618855 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.722049 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.722128 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.722149 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.722193 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.722217 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.825007 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.825082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.825107 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.825137 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.825173 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.928219 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.928258 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.928268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.928285 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.928295 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:01Z","lastTransitionTime":"2026-01-29T15:02:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:01 crc kubenswrapper[4767]: I0129 15:02:01.978135 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 14:21:44.056938942 +0000 UTC Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.017976 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.018033 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.018176 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.018304 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.030797 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.030840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.030850 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.030866 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.030877 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.134119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.134167 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.134179 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.134196 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.134207 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.236565 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.236623 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.236634 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.236654 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.236666 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.339810 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.339847 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.339855 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.339869 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.339880 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.442793 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.442877 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.442937 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.442970 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.442997 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.545566 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.545608 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.545621 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.545636 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.545647 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.649194 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.649243 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.649259 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.649279 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.649295 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.753034 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.753110 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.753126 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.753151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.753168 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.856061 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.856130 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.856148 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.856173 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.856191 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.905811 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.905849 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.905857 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.905869 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.905877 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.919884 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:02Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.923387 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.923412 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.923424 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.923438 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.923449 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.940393 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:02Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.944161 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.944210 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.944222 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.944268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.944281 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.959542 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:02Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.962818 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.962855 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.962867 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.962891 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.962932 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.975650 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:02Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.978316 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 13:46:09.38422993 +0000 UTC Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.979229 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.979268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.979277 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.979292 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.979301 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.992999 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:02Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:02 crc kubenswrapper[4767]: E0129 15:02:02.993242 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.995058 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.995094 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.995103 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.995117 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:02 crc kubenswrapper[4767]: I0129 15:02:02.995128 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:02Z","lastTransitionTime":"2026-01-29T15:02:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.018582 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.018623 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:03 crc kubenswrapper[4767]: E0129 15:02:03.018786 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:03 crc kubenswrapper[4767]: E0129 15:02:03.018870 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.097490 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.097548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.097558 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.097578 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.097590 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.199829 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.199863 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.199871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.199884 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.199907 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.302169 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.302228 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.302238 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.302254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.302262 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.404042 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.404105 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.404123 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.404150 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.404171 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.506912 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.506972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.506992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.507016 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.507045 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.610484 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.610548 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.610566 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.610591 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.610611 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.714048 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.714135 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.714158 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.714191 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.714215 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.817647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.817685 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.817693 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.817709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.817718 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.920074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.920110 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.920118 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.920133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.920142 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:03Z","lastTransitionTime":"2026-01-29T15:02:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:03 crc kubenswrapper[4767]: I0129 15:02:03.979427 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 02:29:31.498039894 +0000 UTC Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.017570 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:04 crc kubenswrapper[4767]: E0129 15:02:04.017814 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.017586 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:04 crc kubenswrapper[4767]: E0129 15:02:04.018003 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.022397 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.022442 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.022452 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.022467 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.022477 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.125188 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.125604 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.125808 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.126066 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.126297 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.229055 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.229092 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.229100 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.229111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.229120 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.332191 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.332508 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.332543 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.332580 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.332604 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.435784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.435841 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.435850 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.435864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.435876 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.539175 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.539238 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.539255 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.539278 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.539294 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.641873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.642225 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.642363 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.642546 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.642694 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.745218 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.745286 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.745308 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.745336 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.745356 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.848021 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.848111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.848129 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.848156 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.848173 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.950992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.951066 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.951087 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.951111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.951132 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:04Z","lastTransitionTime":"2026-01-29T15:02:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:04 crc kubenswrapper[4767]: I0129 15:02:04.980489 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 02:24:13.455549667 +0000 UTC Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.017982 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.018074 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:05 crc kubenswrapper[4767]: E0129 15:02:05.018639 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:05 crc kubenswrapper[4767]: E0129 15:02:05.018841 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.019124 4767 scope.go:117] "RemoveContainer" containerID="3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.037169 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35bc0380-e229-40aa-ab3c-cedf5ee56220\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07aa7ed0ef0aa928fdd9985e1452a58f70b50129ce206e2e423a98b691f841e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.054494 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.054559 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.054579 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.054629 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.054648 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.074315 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.090840 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.109179 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.130551 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.147666 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.164112 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.164166 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.164188 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.164210 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.164224 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.165664 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.181827 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.201356 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"2026-01-29T15:01:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f\\\\n2026-01-29T15:01:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f to /host/opt/cni/bin/\\\\n2026-01-29T15:01:13Z [verbose] multus-daemon started\\\\n2026-01-29T15:01:13Z [verbose] Readiness Indicator file check\\\\n2026-01-29T15:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.218250 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.231281 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.242478 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.257380 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.266946 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.266977 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.266990 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.267007 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.267019 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.269122 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.283503 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.295919 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.305121 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.314209 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.324478 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.369926 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.369970 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.369984 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.370023 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.370052 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.472164 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.472203 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.472212 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.472226 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.472236 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.575078 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.575147 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.575165 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.575191 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.575211 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.637998 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/2.log" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.642082 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.642710 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.660531 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.677391 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.677972 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.678027 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.678044 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.678062 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.678074 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.692229 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.705450 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"2026-01-29T15:01:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f\\\\n2026-01-29T15:01:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f to /host/opt/cni/bin/\\\\n2026-01-29T15:01:13Z [verbose] multus-daemon started\\\\n2026-01-29T15:01:13Z [verbose] Readiness Indicator file check\\\\n2026-01-29T15:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.725172 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.740455 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.753968 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.775417 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.779983 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.780034 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.780045 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.780062 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.780074 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.790751 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.809867 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.824562 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.834105 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.843241 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.853104 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.863974 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35bc0380-e229-40aa-ab3c-cedf5ee56220\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07aa7ed0ef0aa928fdd9985e1452a58f70b50129ce206e2e423a98b691f841e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.883026 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.883085 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.883105 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.883128 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.883145 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.889740 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.904685 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.919290 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.935360 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:02:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:05Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.981172 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 11:41:30.155807073 +0000 UTC Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.985958 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.985992 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.986001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.986014 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:05 crc kubenswrapper[4767]: I0129 15:02:05.986023 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:05Z","lastTransitionTime":"2026-01-29T15:02:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.017833 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.017927 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:06 crc kubenswrapper[4767]: E0129 15:02:06.017978 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:06 crc kubenswrapper[4767]: E0129 15:02:06.018047 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.088119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.088195 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.088209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.088230 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.088242 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.190115 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.190167 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.190177 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.190190 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.190198 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.292528 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.292612 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.292630 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.292655 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.292672 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.395733 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.395801 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.395822 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.395846 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.395864 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.499445 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.499535 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.499552 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.499576 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.499592 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.603010 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.603098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.603120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.603146 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.603168 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.648602 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/3.log" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.649654 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/2.log" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.654678 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" exitCode=1 Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.654757 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.654822 4767 scope.go:117] "RemoveContainer" containerID="3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.656247 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:02:06 crc kubenswrapper[4767]: E0129 15:02:06.656581 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.671044 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.687808 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.704963 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.706518 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.706543 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.706554 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.706569 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.706580 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.719202 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.731791 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.751616 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.764887 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.778648 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35bc0380-e229-40aa-ab3c-cedf5ee56220\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07aa7ed0ef0aa928fdd9985e1452a58f70b50129ce206e2e423a98b691f841e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.798418 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.809403 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.809448 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.809459 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.809476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.809488 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.811862 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.825151 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.843023 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3df9daa5ceacfcc5e93923579afa043f2674230fd0eced2685eac72e406d6089\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:36Z\\\",\\\"message\\\":\\\"4 6413 ovn.go:134] Ensuring zone local for Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk in node crc\\\\nI0129 15:01:36.869709 6413 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-node-tg5zk after 0 failed attempt(s)\\\\nI0129 15:01:36.869714 6413 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-node-tg5zk\\\\nI0129 15:01:36.869726 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0129 15:01:36.869696 6413 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/machine-api-operator-webhook]} name:Service_openshift-machine-api/machine-api-operator-webhook_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.254:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e4e4203e-87c7-4024-930a-5d6bdfe2bdde}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0129 15:01:36.869778 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:02:05Z\\\",\\\"message\\\":\\\"tocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0129 15:02:05.896304 6816 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for network=default: []services.LB{}\\\\nF0129 15:02:05.896314 6816 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:02:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.860570 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.878377 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.894322 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.911443 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"2026-01-29T15:01:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f\\\\n2026-01-29T15:01:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f to /host/opt/cni/bin/\\\\n2026-01-29T15:01:13Z [verbose] multus-daemon started\\\\n2026-01-29T15:01:13Z [verbose] Readiness Indicator file check\\\\n2026-01-29T15:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.911864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.911930 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.911943 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.911960 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.911973 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:06Z","lastTransitionTime":"2026-01-29T15:02:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.933656 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.948996 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.968232 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:06Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:06 crc kubenswrapper[4767]: I0129 15:02:06.982291 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 13:12:40.037582101 +0000 UTC Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.015650 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.015712 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.015737 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.015767 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.015785 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.018250 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.018285 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:07 crc kubenswrapper[4767]: E0129 15:02:07.018427 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:07 crc kubenswrapper[4767]: E0129 15:02:07.018549 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.118962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.119318 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.119465 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.119609 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.119756 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.222695 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.222754 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.222771 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.222793 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.222809 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.325808 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.326181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.326326 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.326463 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.326602 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.430818 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.431208 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.431402 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.431536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.431724 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.535505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.535979 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.536135 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.536289 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.536519 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.639653 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.639718 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.639754 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.639783 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.639803 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.661502 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/3.log" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.667238 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:02:07 crc kubenswrapper[4767]: E0129 15:02:07.667673 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.690598 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.713041 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.726442 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.742931 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"2026-01-29T15:01:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f\\\\n2026-01-29T15:01:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f to /host/opt/cni/bin/\\\\n2026-01-29T15:01:13Z [verbose] multus-daemon started\\\\n2026-01-29T15:01:13Z [verbose] Readiness Indicator file check\\\\n2026-01-29T15:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.744775 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.744847 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.744865 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.744884 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.744919 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.760005 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.774306 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.788354 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.806322 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.821768 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.838522 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.847231 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.847269 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.847284 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.847301 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.847312 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.854385 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.873386 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.887726 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.899989 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.913879 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35bc0380-e229-40aa-ab3c-cedf5ee56220\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07aa7ed0ef0aa928fdd9985e1452a58f70b50129ce206e2e423a98b691f841e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.939698 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.950932 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.950980 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.951019 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.951039 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.951050 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:07Z","lastTransitionTime":"2026-01-29T15:02:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.956220 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.974810 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:07 crc kubenswrapper[4767]: I0129 15:02:07.982513 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 18:14:56.195897074 +0000 UTC Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.000189 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:02:05Z\\\",\\\"message\\\":\\\"tocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0129 15:02:05.896304 6816 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for network=default: []services.LB{}\\\\nF0129 15:02:05.896314 6816 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:02:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:07Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.018155 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:08 crc kubenswrapper[4767]: E0129 15:02:08.018331 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.018797 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:08 crc kubenswrapper[4767]: E0129 15:02:08.018940 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.053904 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.053942 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.053951 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.053976 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.053988 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.157209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.157271 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.157289 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.157324 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.157341 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.261446 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.261491 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.261505 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.261523 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.261535 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.364976 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.365303 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.365397 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.365482 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.365596 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.468739 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.468827 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.468867 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.468930 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.468948 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.572367 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.572678 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.572871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.573059 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.573189 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.676946 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.677025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.677043 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.677066 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.677087 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.782222 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.782273 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.782301 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.782328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.782354 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.885068 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.885194 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.885213 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.885237 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.885253 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.983078 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 18:35:25.97893813 +0000 UTC Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.988460 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.988517 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.988533 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.988555 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:08 crc kubenswrapper[4767]: I0129 15:02:08.988571 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:08Z","lastTransitionTime":"2026-01-29T15:02:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.018560 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.018712 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.018781 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.019031 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.091809 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.091863 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.091880 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.091945 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.091969 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.194682 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.194748 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.194764 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.194787 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.194805 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.298428 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.298873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.299051 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.299257 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.299438 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.402043 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.402429 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.402588 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.402737 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.402875 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.506254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.506636 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.506796 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.506964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.507099 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.610753 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.611118 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.611318 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.611588 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.611820 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.715846 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.715944 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.715964 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.716029 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.716048 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.754310 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:13.754253804 +0000 UTC m=+149.564570883 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.754391 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.755030 4767 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.755135 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:13.755115968 +0000 UTC m=+149.565433037 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.754763 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.818700 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.818774 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.818796 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.818826 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.818845 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.857094 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.857172 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.857223 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857358 4767 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857429 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:13.85741083 +0000 UTC m=+149.667727879 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857535 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857558 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857590 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857612 4767 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857623 4767 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857630 4767 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857709 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:13.857677197 +0000 UTC m=+149.667994266 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:02:09 crc kubenswrapper[4767]: E0129 15:02:09.857736 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:13.857724199 +0000 UTC m=+149.668041268 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.921501 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.921580 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.921597 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.921621 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.921638 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:09Z","lastTransitionTime":"2026-01-29T15:02:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:09 crc kubenswrapper[4767]: I0129 15:02:09.984270 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 09:59:24.333294726 +0000 UTC Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.018155 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.018182 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:10 crc kubenswrapper[4767]: E0129 15:02:10.018375 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:10 crc kubenswrapper[4767]: E0129 15:02:10.018513 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.024813 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.024857 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.024868 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.024885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.024914 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.128323 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.128391 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.128408 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.128435 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.128452 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.231787 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.231837 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.231855 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.231879 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.231919 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.334694 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.334742 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.334758 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.334782 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.334801 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.437482 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.437522 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.437529 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.437543 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.437551 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.540272 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.540314 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.540327 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.540342 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.540353 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.642951 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.643002 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.643010 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.643025 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.643034 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.745262 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.745319 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.745336 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.745357 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.745373 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.847934 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.847994 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.848011 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.848035 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.848056 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.950864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.950978 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.951001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.951064 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.951082 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:10Z","lastTransitionTime":"2026-01-29T15:02:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:10 crc kubenswrapper[4767]: I0129 15:02:10.985308 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 18:25:50.634099803 +0000 UTC Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.018077 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.018146 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:11 crc kubenswrapper[4767]: E0129 15:02:11.018294 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:11 crc kubenswrapper[4767]: E0129 15:02:11.018426 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.054510 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.054585 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.054610 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.054643 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.054669 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.157837 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.157936 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.157954 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.157982 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.157998 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.261022 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.261477 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.261603 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.261701 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.261784 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.365586 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.365839 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.365946 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.366042 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.366120 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.469119 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.469197 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.469209 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.469227 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.469265 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.572122 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.572242 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.572281 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.572315 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.572340 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.676325 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.676391 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.676414 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.676437 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.676451 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.780059 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.780111 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.780128 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.780151 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.780167 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.882623 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.882751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.882771 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.882798 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.882822 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.985402 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 00:20:29.93193089 +0000 UTC Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.986191 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.986233 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.986251 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.986280 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:11 crc kubenswrapper[4767]: I0129 15:02:11.986290 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:11Z","lastTransitionTime":"2026-01-29T15:02:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.018293 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.018320 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:12 crc kubenswrapper[4767]: E0129 15:02:12.018430 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:12 crc kubenswrapper[4767]: E0129 15:02:12.018557 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.088530 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.088638 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.088650 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.088665 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.088675 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.190967 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.191064 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.191100 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.191133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.191155 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.293804 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.293852 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.293863 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.293878 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.293888 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.397338 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.397400 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.397415 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.397439 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.397454 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.500597 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.500663 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.500691 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.500725 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.500747 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.603504 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.603564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.603580 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.603604 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.603620 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.705716 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.705780 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.705797 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.705822 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.705860 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.810030 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.810082 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.810104 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.810126 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.810140 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.913424 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.913476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.913490 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.913509 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.913525 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:12Z","lastTransitionTime":"2026-01-29T15:02:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:12 crc kubenswrapper[4767]: I0129 15:02:12.986600 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 00:33:35.635230152 +0000 UTC Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.016261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.016311 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.016324 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.016341 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.016352 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.019134 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.019177 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.019283 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.019388 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.118950 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.119006 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.119019 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.119038 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.119053 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.212569 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.212624 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.212633 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.212647 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.212657 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.223635 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.226947 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.226977 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.226996 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.227020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.227060 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.240042 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.243417 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.243443 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.243451 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.243466 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.243475 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.257525 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.260949 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.260978 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.260986 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.260997 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.261006 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.272624 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.275466 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.275507 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.275516 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.275531 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.275541 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.294342 4767 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-29T15:02:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d5052518-034d-49f7-b53f-ad999adcfc00\\\",\\\"systemUUID\\\":\\\"1099683a-5760-459d-826c-e4f968e7100f\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:13Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:13 crc kubenswrapper[4767]: E0129 15:02:13.294703 4767 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.297579 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.297737 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.297751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.297782 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.297794 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.400059 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.400122 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.400144 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.400171 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.400192 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.503178 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.503240 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.503261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.503291 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.503310 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.605393 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.605433 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.605443 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.605459 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.605470 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.708183 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.708249 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.708267 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.708299 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.708325 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.812426 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.812544 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.812569 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.812600 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.812621 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.916281 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.916338 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.916354 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.916376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.916392 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:13Z","lastTransitionTime":"2026-01-29T15:02:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:13 crc kubenswrapper[4767]: I0129 15:02:13.987784 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 17:18:57.455956395 +0000 UTC Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.017915 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.017994 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:14 crc kubenswrapper[4767]: E0129 15:02:14.018092 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:14 crc kubenswrapper[4767]: E0129 15:02:14.018202 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.019984 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.020052 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.020074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.020099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.020121 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.123511 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.123551 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.123565 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.123582 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.123592 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.227232 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.227343 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.227361 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.227387 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.227405 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.330010 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.330098 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.330121 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.330179 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.330202 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.433214 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.433293 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.433302 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.433314 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.433323 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.536423 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.536471 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.536508 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.536526 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.536540 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.639564 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.639649 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.639710 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.639739 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.639792 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.742291 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.742344 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.742353 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.742366 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.742376 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.844864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.844962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.844981 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.845007 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.845025 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.952482 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.952532 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.952544 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.952560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.952572 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:14Z","lastTransitionTime":"2026-01-29T15:02:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:14 crc kubenswrapper[4767]: I0129 15:02:14.988451 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 14:11:59.346446278 +0000 UTC Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.019155 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.019218 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:15 crc kubenswrapper[4767]: E0129 15:02:15.019322 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:15 crc kubenswrapper[4767]: E0129 15:02:15.019597 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.031600 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad5d63c18b23b7596a3c0bab575e6ab96307d53716c32e830858c77c60c57cea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.042756 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zqvls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3395d14-43f2-45f5-8692-034412aff69e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08590aa9306b6d01e0aac37b1bef4e2cf0e253fc6a3d07ee5d1d9d8a6efc3010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bfv7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zqvls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.052552 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03807b98-087e-4b91-b3ea-f2457587c580\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mkl5w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:24Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-lhmrn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.054560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.054670 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.054691 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.054752 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.054772 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.064583 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b3376c25-a1eb-41bc-95c8-14a6f62d5b4a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19cf5b0ff7c4c84aa1dfb965ab775f23c4b842e593e086ac3406e3f88a029b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://193392d1610fd2c665c8010e1a51381c983b60a753d62036bfa776f983aec2c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df584f9048180a53178e560bf9b45975541235e823e79df02387392e71e0ebd4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.075988 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6b7f16-23dc-4d41-aa16-8c05a62b9a1c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3041dafbc61bc8efb6ec501a6851ddc7a8ec95bdf3c428c5fcba1041335ef1ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://12c79feb4d06225fb805bab15c142816b48b15d41c2346a730183d1a82da5b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0aec999e3f64a14b9a9769e5f9a4f3d9af8ed849658bb0cb552440d06a210f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff6585182a096e96eb22d24f667a4e79d047dade2a6b56688fd195e342015dc0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.091032 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.107461 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.117882 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c144460d-d956-4e9c-8354-bfd72b970e30\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c12bea7f94306ac2c8babab9ca424d8f99eb8e849e96414bbf6d758bf78ca36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5hbb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kgsln\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.134292 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35bc0380-e229-40aa-ab3c-cedf5ee56220\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d07aa7ed0ef0aa928fdd9985e1452a58f70b50129ce206e2e423a98b691f841e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c15cce530fab94d205d755a035559d46765ed41c6e65bb4950e5db35407da91\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.156661 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.156697 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.156709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.156729 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.156742 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.162010 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4db36907-470a-4662-8f5f-04b7667100ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c65136db8a0e9d2b6f00076b6528793746a283c2aa88d9db099cd83b0adc103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d0aef4649578efe2549dd87c86b8d6f0fe5b49b3fb3bb66496a755680ec3f953\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f710bc6525bde0da7d7390539753ea8f7fe0b51b16747608ad201c1240014c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb5046a08558e5b111245029a7a7549932b0ea03f54f7ff331e027104d74109e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28d3557d0d70a885b16d8dabcf149c5a336529a332dca123e37e385437c9e10\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://179649679bf8f8a42460995ad15bf6306dbbca7bfe5b05477ce0fe86d898a06c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6213f20504d5ad115a89b321ffaffade0b636ba67a716ccc7b01dfff34af1f7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c8b8829b5dba099d55c35d409e8bce4830f7f7c5517612529b2919838ee8e63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.176995 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.194499 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://943e611ac51892386e877e88149e41ddb1fb3eee83d9321088aa00e7d559cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18fd4561ca344ee0005543fe574a17f3b892b6f2dbbc0c35e31576a564395e59\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.221949 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"633e55cd-6740-4d13-84ef-023b691c0428\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:02:05Z\\\",\\\"message\\\":\\\"tocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.140\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0129 15:02:05.896304 6816 services_controller.go:452] Built service openshift-marketplace/redhat-marketplace per-node LB for network=default: []services.LB{}\\\\nF0129 15:02:05.896314 6816 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:02:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dw776\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-tg5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.279824 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c02de73-c364-474f-b2ff-86ed8e52530a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c25704951987b888b062e82a284abb79d5a0507e3c6e7d2862cef65d33f8f2db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2d782a43a9a34163578080b0d6e8debfb3772ed77292612ce116cb5edc7acb3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7257m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-qmhpr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.281346 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.281461 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.281475 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.281496 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.281512 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.299451 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"45069754-41c6-428c-bef5-149c235f5944\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:00:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-29T15:01:05Z\\\",\\\"message\\\":\\\"file observer\\\\nW0129 15:01:05.066436 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0129 15:01:05.066585 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0129 15:01:05.067488 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-997310900/tls.crt::/tmp/serving-cert-997310900/tls.key\\\\\\\"\\\\nI0129 15:01:05.359860 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0129 15:01:05.364243 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0129 15:01:05.364264 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0129 15:01:05.364285 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0129 15:01:05.364289 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0129 15:01:05.369990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0129 15:01:05.370007 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0129 15:01:05.370031 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370042 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0129 15:01:05.370053 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0129 15:01:05.370060 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0129 15:01:05.370067 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0129 15:01:05.370073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0129 15:01:05.371622 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:00:48Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:00:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:00:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:00:45Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.312244 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:07Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf71374be1c5827f8c3c7d49b1d105143d42d1857ebf21bc3858e496016500fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.321958 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-k6xcf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afd48dc4-70ed-45d0-8133-5a6f2ef7014f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://caa167275c5a6fde5af0c377d2d08c998aecb4c84779723efd54556225e56942\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-98h48\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-k6xcf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.334475 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-9xcr9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4615231a-e157-430e-9ecc-97c3fd7701bb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-29T15:01:58Z\\\",\\\"message\\\":\\\"2026-01-29T15:01:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f\\\\n2026-01-29T15:01:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_17c4a049-60ee-478e-8ddc-adf0ab00a55f to /host/opt/cni/bin/\\\\n2026-01-29T15:01:13Z [verbose] multus-daemon started\\\\n2026-01-29T15:01:13Z [verbose] Readiness Indicator file check\\\\n2026-01-29T15:01:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mcp9z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-9xcr9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.349240 4767 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"65812f82-4e2a-4f3d-bf4c-3aa608f3de8d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-29T15:01:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://124290b31a3892b268a53e69f45cf4b9cc4503e1e95675c9a4d691787a6c4123\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-29T15:01:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99e8b1a885c8e4ef0d66718858915db243f1b5329ef3de19794cbe6cdfabd1e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4f9ad0bd88edfab01a0d546576b61534b9b2194bd4764b11813a55d6e65c826\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a9746b18419c3fed817e3f95aced38889507cf02b1b58266f0d878e38edbde94\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ac8b5541e103541f8dfd7859c9f9dbccb7aaaf25cb5619a90725c3061b096ae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://41c27d7f87ab255a8f5b7ef554b24d6bcc9e370bab110460760c10b89e0228cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c4103c238078a0c7c4481e3f42b844473167e960e2328f5a8fe2a7578461456\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-29T15:01:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-29T15:01:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cj6dz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-29T15:01:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-2qt9f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-29T15:02:15Z is after 2025-08-24T17:21:41Z" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.384533 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.384590 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.384606 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.384624 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.384636 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.487353 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.487430 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.487449 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.487470 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.487486 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.590133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.590167 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.590179 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.590194 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.590206 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.692575 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.692632 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.692671 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.692693 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.692710 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.796224 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.796281 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.796299 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.796321 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.796337 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.898799 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.898873 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.898890 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.898952 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.898970 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:15Z","lastTransitionTime":"2026-01-29T15:02:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:15 crc kubenswrapper[4767]: I0129 15:02:15.988537 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 16:55:35.741744087 +0000 UTC Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.001556 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.001634 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.001653 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.001677 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.001695 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.017939 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.017976 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:16 crc kubenswrapper[4767]: E0129 15:02:16.018146 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:16 crc kubenswrapper[4767]: E0129 15:02:16.018245 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.104763 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.104801 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.104809 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.104829 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.104838 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.208133 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.208394 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.208413 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.208436 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.208453 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.310947 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.311027 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.311041 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.311061 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.311077 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.414227 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.414318 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.414337 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.414363 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.414382 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.517639 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.517711 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.517729 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.517751 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.517769 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.620015 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.620060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.620071 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.620089 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.620101 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.722320 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.722364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.722375 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.722390 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.722403 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.825745 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.825810 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.825823 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.825840 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.825852 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.929088 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.929158 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.929172 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.929190 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.929202 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:16Z","lastTransitionTime":"2026-01-29T15:02:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:16 crc kubenswrapper[4767]: I0129 15:02:16.988976 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 15:38:39.430758808 +0000 UTC Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.018503 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.018506 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:17 crc kubenswrapper[4767]: E0129 15:02:17.018717 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:17 crc kubenswrapper[4767]: E0129 15:02:17.019148 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.032394 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.032481 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.032499 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.032521 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.032539 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.136086 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.136164 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.136181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.136205 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.136223 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.239536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.239601 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.239619 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.239642 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.239659 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.342686 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.342746 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.342768 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.342796 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.342820 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.445637 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.445681 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.445690 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.445702 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.445712 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.549192 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.549254 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.549268 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.549290 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.549305 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.652202 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.652288 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.652307 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.652333 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.652351 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.755709 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.755790 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.755810 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.755834 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.755852 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.859247 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.859335 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.859364 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.859396 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.859416 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.962639 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.962673 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.962682 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.962694 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.962702 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:17Z","lastTransitionTime":"2026-01-29T15:02:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:17 crc kubenswrapper[4767]: I0129 15:02:17.989423 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 22:42:50.61609809 +0000 UTC Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.018440 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.018520 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:18 crc kubenswrapper[4767]: E0129 15:02:18.018756 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:18 crc kubenswrapper[4767]: E0129 15:02:18.018991 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.064934 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.064968 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.064976 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.064990 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.065000 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.167406 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.167461 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.167475 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.167494 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.167512 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.270284 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.270360 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.270377 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.270405 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.270423 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.373818 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.373945 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.373981 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.374009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.374025 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.477997 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.478054 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.478070 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.478095 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.478112 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.581203 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.581274 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.581308 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.581342 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.581362 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.684177 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.684272 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.684297 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.684326 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.684352 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.787644 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.787725 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.787749 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.787771 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.787787 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.890445 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.890524 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.890543 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.890568 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.890587 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.990118 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 17:08:59.295116355 +0000 UTC Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.993433 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.993488 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.993523 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.993560 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:18 crc kubenswrapper[4767]: I0129 15:02:18.993582 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:18Z","lastTransitionTime":"2026-01-29T15:02:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.018149 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:19 crc kubenswrapper[4767]: E0129 15:02:19.018392 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.018455 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:19 crc kubenswrapper[4767]: E0129 15:02:19.018849 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.096846 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.097158 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.097170 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.097187 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.097201 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.200095 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.200147 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.200164 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.200187 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.200219 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.303044 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.303120 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.303144 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.303172 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.303195 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.407093 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.407172 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.407202 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.407233 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.407255 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.510661 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.510733 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.510755 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.510778 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.510797 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.613416 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.613501 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.613542 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.613573 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.613594 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.716779 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.716843 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.716864 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.716885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.716925 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.819449 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.819497 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.819507 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.819523 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.819534 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.923476 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.923543 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.923578 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.923610 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.923630 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:19Z","lastTransitionTime":"2026-01-29T15:02:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:19 crc kubenswrapper[4767]: I0129 15:02:19.991324 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 04:11:04.20078773 +0000 UTC Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.017940 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.017954 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:20 crc kubenswrapper[4767]: E0129 15:02:20.018169 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:20 crc kubenswrapper[4767]: E0129 15:02:20.018242 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.026611 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.026654 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.026712 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.026734 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.026750 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.130313 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.130381 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.130405 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.130435 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.130453 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.233697 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.233764 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.233784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.233807 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.233825 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.337513 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.337577 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.337595 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.337619 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.337637 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.439719 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.439797 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.439820 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.439848 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.439869 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.543013 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.543078 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.543095 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.543118 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.543136 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.646772 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.646838 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.646855 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.646885 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.646939 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.750157 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.750218 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.750235 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.750258 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.750277 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.852843 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.852932 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.852947 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.852963 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.852975 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.956572 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.956652 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.956675 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.956703 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.956720 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:20Z","lastTransitionTime":"2026-01-29T15:02:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:20 crc kubenswrapper[4767]: I0129 15:02:20.992183 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 18:27:34.355070672 +0000 UTC Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.017955 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.017991 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:21 crc kubenswrapper[4767]: E0129 15:02:21.018365 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:21 crc kubenswrapper[4767]: E0129 15:02:21.018502 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.059035 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.059099 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.059117 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.059145 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.059182 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.161642 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.161704 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.161717 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.161742 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.161757 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.265201 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.265286 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.265305 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.265331 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.265350 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.368500 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.368546 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.368557 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.368576 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.368588 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.471856 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.471984 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.472004 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.472032 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.472050 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.575847 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.575980 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.576005 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.576039 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.576063 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.679126 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.679169 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.679181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.679194 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.679204 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.781846 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.781890 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.781921 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.781936 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.781946 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.885249 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.885297 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.885309 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.885328 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.885340 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.990727 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.990784 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.990798 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.990826 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.990839 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:21Z","lastTransitionTime":"2026-01-29T15:02:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:21 crc kubenswrapper[4767]: I0129 15:02:21.992280 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 17:28:12.44324045 +0000 UTC Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.017690 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.017782 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:22 crc kubenswrapper[4767]: E0129 15:02:22.017949 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:22 crc kubenswrapper[4767]: E0129 15:02:22.018545 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.018994 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:02:22 crc kubenswrapper[4767]: E0129 15:02:22.019243 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.094419 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.094879 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.095156 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.095336 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.095474 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.198811 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.198871 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.198888 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.198938 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.198956 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.302693 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.302722 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.302730 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.302744 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.302753 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.405740 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.405779 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.405792 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.405808 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.405820 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.508682 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.508755 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.508780 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.508814 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.508838 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.612717 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.612770 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.612786 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.612806 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.612821 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.714881 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.714962 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.714979 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.715003 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.715020 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.818074 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.818135 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.818152 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.818181 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.818204 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.921315 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.921376 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.921392 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.921415 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.921432 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:22Z","lastTransitionTime":"2026-01-29T15:02:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:22 crc kubenswrapper[4767]: I0129 15:02:22.993965 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 18:42:01.128677473 +0000 UTC Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.017719 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:23 crc kubenswrapper[4767]: E0129 15:02:23.017889 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.018023 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:23 crc kubenswrapper[4767]: E0129 15:02:23.018286 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.024446 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.024500 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.024519 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.024536 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.024549 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:23Z","lastTransitionTime":"2026-01-29T15:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.126961 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.127009 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.127020 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.127036 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.127047 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:23Z","lastTransitionTime":"2026-01-29T15:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.232021 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.232060 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.232068 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.232081 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.232090 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:23Z","lastTransitionTime":"2026-01-29T15:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.334780 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.334819 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.334829 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.334843 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.334853 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:23Z","lastTransitionTime":"2026-01-29T15:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.362217 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.362253 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.362261 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.362274 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.362283 4767 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-29T15:02:23Z","lastTransitionTime":"2026-01-29T15:02:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.401621 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4"] Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.402035 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.403640 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.403969 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.404022 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.405653 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.499364 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=76.499342895 podStartE2EDuration="1m16.499342895s" podCreationTimestamp="2026-01-29 15:01:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.499303594 +0000 UTC m=+99.309620643" watchObservedRunningTime="2026-01-29 15:02:23.499342895 +0000 UTC m=+99.309659954" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.499818 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=22.499811788 podStartE2EDuration="22.499811788s" podCreationTimestamp="2026-01-29 15:02:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.474649847 +0000 UTC m=+99.284966906" watchObservedRunningTime="2026-01-29 15:02:23.499811788 +0000 UTC m=+99.310128837" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.517245 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aef4cf0-3989-467d-a15b-06e2b8f9f396-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.517288 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0aef4cf0-3989-467d-a15b-06e2b8f9f396-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.517316 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0aef4cf0-3989-467d-a15b-06e2b8f9f396-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.517333 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0aef4cf0-3989-467d-a15b-06e2b8f9f396-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.517351 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0aef4cf0-3989-467d-a15b-06e2b8f9f396-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.543523 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-9xcr9" podStartSLOduration=73.543502786 podStartE2EDuration="1m13.543502786s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.523757526 +0000 UTC m=+99.334074605" watchObservedRunningTime="2026-01-29 15:02:23.543502786 +0000 UTC m=+99.353819825" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.557201 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-2qt9f" podStartSLOduration=73.557180548 podStartE2EDuration="1m13.557180548s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.544848334 +0000 UTC m=+99.355165393" watchObservedRunningTime="2026-01-29 15:02:23.557180548 +0000 UTC m=+99.367497587" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.572220 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qmhpr" podStartSLOduration=73.572204376 podStartE2EDuration="1m13.572204376s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.557505067 +0000 UTC m=+99.367822106" watchObservedRunningTime="2026-01-29 15:02:23.572204376 +0000 UTC m=+99.382521415" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.585006 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=77.584986263 podStartE2EDuration="1m17.584986263s" podCreationTimestamp="2026-01-29 15:01:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.57270692 +0000 UTC m=+99.383023959" watchObservedRunningTime="2026-01-29 15:02:23.584986263 +0000 UTC m=+99.395303302" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.607193 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-k6xcf" podStartSLOduration=73.607174511 podStartE2EDuration="1m13.607174511s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.597151832 +0000 UTC m=+99.407468871" watchObservedRunningTime="2026-01-29 15:02:23.607174511 +0000 UTC m=+99.417491550" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.618118 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aef4cf0-3989-467d-a15b-06e2b8f9f396-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.618163 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0aef4cf0-3989-467d-a15b-06e2b8f9f396-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.618197 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0aef4cf0-3989-467d-a15b-06e2b8f9f396-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.618222 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0aef4cf0-3989-467d-a15b-06e2b8f9f396-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.618245 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0aef4cf0-3989-467d-a15b-06e2b8f9f396-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.618308 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0aef4cf0-3989-467d-a15b-06e2b8f9f396-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.618318 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0aef4cf0-3989-467d-a15b-06e2b8f9f396-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.619146 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0aef4cf0-3989-467d-a15b-06e2b8f9f396-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.624649 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0aef4cf0-3989-467d-a15b-06e2b8f9f396-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.633757 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podStartSLOduration=73.633741772 podStartE2EDuration="1m13.633741772s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.633619738 +0000 UTC m=+99.443936777" watchObservedRunningTime="2026-01-29 15:02:23.633741772 +0000 UTC m=+99.444058811" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.637362 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0aef4cf0-3989-467d-a15b-06e2b8f9f396-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dwm4\" (UID: \"0aef4cf0-3989-467d-a15b-06e2b8f9f396\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.652506 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zqvls" podStartSLOduration=73.652491594 podStartE2EDuration="1m13.652491594s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.643044031 +0000 UTC m=+99.453361070" watchObservedRunningTime="2026-01-29 15:02:23.652491594 +0000 UTC m=+99.462808633" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.677863 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.67781058 podStartE2EDuration="1m18.67781058s" podCreationTimestamp="2026-01-29 15:01:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.677297156 +0000 UTC m=+99.487614195" watchObservedRunningTime="2026-01-29 15:02:23.67781058 +0000 UTC m=+99.488127619" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.689466 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=50.689429584 podStartE2EDuration="50.689429584s" podCreationTimestamp="2026-01-29 15:01:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:23.689062994 +0000 UTC m=+99.499380043" watchObservedRunningTime="2026-01-29 15:02:23.689429584 +0000 UTC m=+99.499746643" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.733643 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.995076 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 02:36:44.662207706 +0000 UTC Jan 29 15:02:23 crc kubenswrapper[4767]: I0129 15:02:23.995525 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 29 15:02:24 crc kubenswrapper[4767]: I0129 15:02:24.004037 4767 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 29 15:02:24 crc kubenswrapper[4767]: I0129 15:02:24.018265 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:24 crc kubenswrapper[4767]: I0129 15:02:24.018320 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:24 crc kubenswrapper[4767]: E0129 15:02:24.018387 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:24 crc kubenswrapper[4767]: E0129 15:02:24.018469 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:24 crc kubenswrapper[4767]: I0129 15:02:24.722406 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" event={"ID":"0aef4cf0-3989-467d-a15b-06e2b8f9f396","Type":"ContainerStarted","Data":"90e16bb61a76a07ceb5719b2cbef12eccb89584cce2f9bd8ac8e625f57936627"} Jan 29 15:02:24 crc kubenswrapper[4767]: I0129 15:02:24.722451 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" event={"ID":"0aef4cf0-3989-467d-a15b-06e2b8f9f396","Type":"ContainerStarted","Data":"e14c958e1dfbf04d42b9acf1983e6ff859338753cb809507fbb0aa53804337c7"} Jan 29 15:02:24 crc kubenswrapper[4767]: I0129 15:02:24.734460 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dwm4" podStartSLOduration=74.734446334 podStartE2EDuration="1m14.734446334s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:24.733672122 +0000 UTC m=+100.543989161" watchObservedRunningTime="2026-01-29 15:02:24.734446334 +0000 UTC m=+100.544763373" Jan 29 15:02:25 crc kubenswrapper[4767]: I0129 15:02:25.017988 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:25 crc kubenswrapper[4767]: I0129 15:02:25.018079 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:25 crc kubenswrapper[4767]: E0129 15:02:25.019234 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:25 crc kubenswrapper[4767]: E0129 15:02:25.019328 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:26 crc kubenswrapper[4767]: I0129 15:02:26.017701 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:26 crc kubenswrapper[4767]: I0129 15:02:26.017764 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:26 crc kubenswrapper[4767]: E0129 15:02:26.017812 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:26 crc kubenswrapper[4767]: E0129 15:02:26.017876 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:27 crc kubenswrapper[4767]: I0129 15:02:27.018543 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:27 crc kubenswrapper[4767]: I0129 15:02:27.018702 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:27 crc kubenswrapper[4767]: E0129 15:02:27.018836 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:27 crc kubenswrapper[4767]: E0129 15:02:27.018964 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:28 crc kubenswrapper[4767]: I0129 15:02:28.017722 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:28 crc kubenswrapper[4767]: I0129 15:02:28.017743 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:28 crc kubenswrapper[4767]: E0129 15:02:28.017856 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:28 crc kubenswrapper[4767]: E0129 15:02:28.018028 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:28 crc kubenswrapper[4767]: I0129 15:02:28.771185 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:28 crc kubenswrapper[4767]: E0129 15:02:28.771449 4767 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:02:28 crc kubenswrapper[4767]: E0129 15:02:28.771937 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs podName:03807b98-087e-4b91-b3ea-f2457587c580 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:32.771914539 +0000 UTC m=+168.582231578 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs") pod "network-metrics-daemon-lhmrn" (UID: "03807b98-087e-4b91-b3ea-f2457587c580") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 29 15:02:29 crc kubenswrapper[4767]: I0129 15:02:29.018020 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:29 crc kubenswrapper[4767]: E0129 15:02:29.018183 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:29 crc kubenswrapper[4767]: I0129 15:02:29.018368 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:29 crc kubenswrapper[4767]: E0129 15:02:29.018574 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:30 crc kubenswrapper[4767]: I0129 15:02:30.018160 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:30 crc kubenswrapper[4767]: I0129 15:02:30.018158 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:30 crc kubenswrapper[4767]: E0129 15:02:30.018852 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:30 crc kubenswrapper[4767]: E0129 15:02:30.018985 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:31 crc kubenswrapper[4767]: I0129 15:02:31.018493 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:31 crc kubenswrapper[4767]: E0129 15:02:31.018630 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:31 crc kubenswrapper[4767]: I0129 15:02:31.018744 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:31 crc kubenswrapper[4767]: E0129 15:02:31.018858 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:32 crc kubenswrapper[4767]: I0129 15:02:32.018354 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:32 crc kubenswrapper[4767]: I0129 15:02:32.018355 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:32 crc kubenswrapper[4767]: E0129 15:02:32.018638 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:32 crc kubenswrapper[4767]: E0129 15:02:32.018758 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:33 crc kubenswrapper[4767]: I0129 15:02:33.017814 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:33 crc kubenswrapper[4767]: E0129 15:02:33.017999 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:33 crc kubenswrapper[4767]: I0129 15:02:33.018196 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:33 crc kubenswrapper[4767]: E0129 15:02:33.018337 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:33 crc kubenswrapper[4767]: I0129 15:02:33.019054 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:02:33 crc kubenswrapper[4767]: E0129 15:02:33.019604 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-tg5zk_openshift-ovn-kubernetes(633e55cd-6740-4d13-84ef-023b691c0428)\"" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" Jan 29 15:02:34 crc kubenswrapper[4767]: I0129 15:02:34.018087 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:34 crc kubenswrapper[4767]: I0129 15:02:34.018130 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:34 crc kubenswrapper[4767]: E0129 15:02:34.018219 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:34 crc kubenswrapper[4767]: E0129 15:02:34.018325 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:35 crc kubenswrapper[4767]: I0129 15:02:35.018783 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:35 crc kubenswrapper[4767]: I0129 15:02:35.021016 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:35 crc kubenswrapper[4767]: E0129 15:02:35.021299 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:35 crc kubenswrapper[4767]: E0129 15:02:35.021448 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:36 crc kubenswrapper[4767]: I0129 15:02:36.018324 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:36 crc kubenswrapper[4767]: I0129 15:02:36.018410 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:36 crc kubenswrapper[4767]: E0129 15:02:36.018562 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:36 crc kubenswrapper[4767]: E0129 15:02:36.018692 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:37 crc kubenswrapper[4767]: I0129 15:02:37.017756 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:37 crc kubenswrapper[4767]: E0129 15:02:37.017922 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:37 crc kubenswrapper[4767]: I0129 15:02:37.018178 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:37 crc kubenswrapper[4767]: E0129 15:02:37.018245 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:38 crc kubenswrapper[4767]: I0129 15:02:38.018302 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:38 crc kubenswrapper[4767]: I0129 15:02:38.018328 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:38 crc kubenswrapper[4767]: E0129 15:02:38.018435 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:38 crc kubenswrapper[4767]: E0129 15:02:38.018516 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:39 crc kubenswrapper[4767]: I0129 15:02:39.018495 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:39 crc kubenswrapper[4767]: I0129 15:02:39.018532 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:39 crc kubenswrapper[4767]: E0129 15:02:39.018655 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:39 crc kubenswrapper[4767]: E0129 15:02:39.018755 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:40 crc kubenswrapper[4767]: I0129 15:02:40.017556 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:40 crc kubenswrapper[4767]: E0129 15:02:40.017669 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:40 crc kubenswrapper[4767]: I0129 15:02:40.017575 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:40 crc kubenswrapper[4767]: E0129 15:02:40.017847 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:41 crc kubenswrapper[4767]: I0129 15:02:41.018318 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:41 crc kubenswrapper[4767]: I0129 15:02:41.018318 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:41 crc kubenswrapper[4767]: E0129 15:02:41.018454 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:41 crc kubenswrapper[4767]: E0129 15:02:41.018509 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:42 crc kubenswrapper[4767]: I0129 15:02:42.017507 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:42 crc kubenswrapper[4767]: I0129 15:02:42.017507 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:42 crc kubenswrapper[4767]: E0129 15:02:42.017621 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:42 crc kubenswrapper[4767]: E0129 15:02:42.017680 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:43 crc kubenswrapper[4767]: I0129 15:02:43.018535 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:43 crc kubenswrapper[4767]: I0129 15:02:43.018580 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:43 crc kubenswrapper[4767]: E0129 15:02:43.019417 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:43 crc kubenswrapper[4767]: E0129 15:02:43.019993 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.017913 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:44 crc kubenswrapper[4767]: E0129 15:02:44.018054 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.017927 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:44 crc kubenswrapper[4767]: E0129 15:02:44.018281 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.783175 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/1.log" Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.783927 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/0.log" Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.784024 4767 generic.go:334] "Generic (PLEG): container finished" podID="4615231a-e157-430e-9ecc-97c3fd7701bb" containerID="37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b" exitCode=1 Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.784083 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerDied","Data":"37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b"} Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.784207 4767 scope.go:117] "RemoveContainer" containerID="4c4694920e17b97886e492bd25bdc44c78cd4b8ff5787a87397777e87f0e113d" Jan 29 15:02:44 crc kubenswrapper[4767]: I0129 15:02:44.784608 4767 scope.go:117] "RemoveContainer" containerID="37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b" Jan 29 15:02:44 crc kubenswrapper[4767]: E0129 15:02:44.784927 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-9xcr9_openshift-multus(4615231a-e157-430e-9ecc-97c3fd7701bb)\"" pod="openshift-multus/multus-9xcr9" podUID="4615231a-e157-430e-9ecc-97c3fd7701bb" Jan 29 15:02:45 crc kubenswrapper[4767]: I0129 15:02:45.017679 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:45 crc kubenswrapper[4767]: I0129 15:02:45.017699 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:45 crc kubenswrapper[4767]: E0129 15:02:45.019956 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:45 crc kubenswrapper[4767]: E0129 15:02:45.019803 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:45 crc kubenswrapper[4767]: E0129 15:02:45.029244 4767 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 29 15:02:45 crc kubenswrapper[4767]: E0129 15:02:45.126829 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 15:02:45 crc kubenswrapper[4767]: I0129 15:02:45.789300 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/1.log" Jan 29 15:02:46 crc kubenswrapper[4767]: I0129 15:02:46.017622 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:46 crc kubenswrapper[4767]: E0129 15:02:46.017792 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:46 crc kubenswrapper[4767]: I0129 15:02:46.017635 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:46 crc kubenswrapper[4767]: E0129 15:02:46.018028 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:47 crc kubenswrapper[4767]: I0129 15:02:47.018304 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:47 crc kubenswrapper[4767]: E0129 15:02:47.018435 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:47 crc kubenswrapper[4767]: I0129 15:02:47.018545 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:47 crc kubenswrapper[4767]: E0129 15:02:47.018685 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.017757 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.017797 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:48 crc kubenswrapper[4767]: E0129 15:02:48.017929 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:48 crc kubenswrapper[4767]: E0129 15:02:48.018090 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.018816 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.766480 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lhmrn"] Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.766667 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:48 crc kubenswrapper[4767]: E0129 15:02:48.766781 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.810880 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/3.log" Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.813852 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerStarted","Data":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} Jan 29 15:02:48 crc kubenswrapper[4767]: I0129 15:02:48.814482 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:02:49 crc kubenswrapper[4767]: I0129 15:02:49.018035 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:49 crc kubenswrapper[4767]: E0129 15:02:49.018502 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:50 crc kubenswrapper[4767]: I0129 15:02:50.017881 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:50 crc kubenswrapper[4767]: I0129 15:02:50.017938 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:50 crc kubenswrapper[4767]: E0129 15:02:50.019022 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:50 crc kubenswrapper[4767]: E0129 15:02:50.019086 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:50 crc kubenswrapper[4767]: E0129 15:02:50.129879 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 15:02:51 crc kubenswrapper[4767]: I0129 15:02:51.018526 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:51 crc kubenswrapper[4767]: E0129 15:02:51.018644 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:51 crc kubenswrapper[4767]: I0129 15:02:51.018534 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:51 crc kubenswrapper[4767]: E0129 15:02:51.018822 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:52 crc kubenswrapper[4767]: I0129 15:02:52.017801 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:52 crc kubenswrapper[4767]: I0129 15:02:52.017846 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:52 crc kubenswrapper[4767]: E0129 15:02:52.017993 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:52 crc kubenswrapper[4767]: E0129 15:02:52.018096 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:53 crc kubenswrapper[4767]: I0129 15:02:53.018157 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:53 crc kubenswrapper[4767]: I0129 15:02:53.018157 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:53 crc kubenswrapper[4767]: E0129 15:02:53.018371 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:53 crc kubenswrapper[4767]: E0129 15:02:53.018725 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:54 crc kubenswrapper[4767]: I0129 15:02:54.017486 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:54 crc kubenswrapper[4767]: I0129 15:02:54.017623 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:54 crc kubenswrapper[4767]: E0129 15:02:54.017652 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:54 crc kubenswrapper[4767]: E0129 15:02:54.017819 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:55 crc kubenswrapper[4767]: I0129 15:02:55.018593 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:55 crc kubenswrapper[4767]: I0129 15:02:55.018617 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:55 crc kubenswrapper[4767]: E0129 15:02:55.020269 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:55 crc kubenswrapper[4767]: E0129 15:02:55.020476 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:55 crc kubenswrapper[4767]: E0129 15:02:55.130416 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 15:02:56 crc kubenswrapper[4767]: I0129 15:02:56.018282 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:56 crc kubenswrapper[4767]: I0129 15:02:56.018371 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:56 crc kubenswrapper[4767]: E0129 15:02:56.018506 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:56 crc kubenswrapper[4767]: E0129 15:02:56.018618 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:57 crc kubenswrapper[4767]: I0129 15:02:57.017822 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:57 crc kubenswrapper[4767]: I0129 15:02:57.017904 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:57 crc kubenswrapper[4767]: E0129 15:02:57.018081 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:57 crc kubenswrapper[4767]: E0129 15:02:57.018174 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:58 crc kubenswrapper[4767]: I0129 15:02:58.018507 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:02:58 crc kubenswrapper[4767]: E0129 15:02:58.018644 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:02:58 crc kubenswrapper[4767]: I0129 15:02:58.018863 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:02:58 crc kubenswrapper[4767]: E0129 15:02:58.018955 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:02:59 crc kubenswrapper[4767]: I0129 15:02:59.018435 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:02:59 crc kubenswrapper[4767]: E0129 15:02:59.018640 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:02:59 crc kubenswrapper[4767]: I0129 15:02:59.018842 4767 scope.go:117] "RemoveContainer" containerID="37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b" Jan 29 15:02:59 crc kubenswrapper[4767]: I0129 15:02:59.019164 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:02:59 crc kubenswrapper[4767]: E0129 15:02:59.019406 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:02:59 crc kubenswrapper[4767]: I0129 15:02:59.051605 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podStartSLOduration=109.051584027 podStartE2EDuration="1m49.051584027s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:02:48.847455785 +0000 UTC m=+124.657772824" watchObservedRunningTime="2026-01-29 15:02:59.051584027 +0000 UTC m=+134.861901076" Jan 29 15:02:59 crc kubenswrapper[4767]: I0129 15:02:59.853180 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/1.log" Jan 29 15:02:59 crc kubenswrapper[4767]: I0129 15:02:59.853587 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerStarted","Data":"f1630240852f2539ae4c87bd3e8b6e0a401724521e6bb6888ffe71ccb369f8f3"} Jan 29 15:03:00 crc kubenswrapper[4767]: I0129 15:03:00.017921 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:00 crc kubenswrapper[4767]: I0129 15:03:00.017954 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:03:00 crc kubenswrapper[4767]: E0129 15:03:00.018082 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:03:00 crc kubenswrapper[4767]: E0129 15:03:00.018191 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:03:00 crc kubenswrapper[4767]: E0129 15:03:00.131566 4767 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 29 15:03:01 crc kubenswrapper[4767]: I0129 15:03:01.017751 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:03:01 crc kubenswrapper[4767]: E0129 15:03:01.018012 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:03:01 crc kubenswrapper[4767]: I0129 15:03:01.018203 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:01 crc kubenswrapper[4767]: E0129 15:03:01.018426 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:03:02 crc kubenswrapper[4767]: I0129 15:03:02.017667 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:03:02 crc kubenswrapper[4767]: I0129 15:03:02.017751 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:02 crc kubenswrapper[4767]: E0129 15:03:02.018614 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:03:02 crc kubenswrapper[4767]: E0129 15:03:02.018460 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:03:03 crc kubenswrapper[4767]: I0129 15:03:03.018198 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:03 crc kubenswrapper[4767]: I0129 15:03:03.018283 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:03:03 crc kubenswrapper[4767]: E0129 15:03:03.018381 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:03:03 crc kubenswrapper[4767]: E0129 15:03:03.018541 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:03:04 crc kubenswrapper[4767]: I0129 15:03:04.017632 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:04 crc kubenswrapper[4767]: I0129 15:03:04.017683 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:03:04 crc kubenswrapper[4767]: E0129 15:03:04.017855 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 29 15:03:04 crc kubenswrapper[4767]: E0129 15:03:04.017996 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 29 15:03:05 crc kubenswrapper[4767]: I0129 15:03:05.018348 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:03:05 crc kubenswrapper[4767]: E0129 15:03:05.019340 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-lhmrn" podUID="03807b98-087e-4b91-b3ea-f2457587c580" Jan 29 15:03:05 crc kubenswrapper[4767]: I0129 15:03:05.019391 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:05 crc kubenswrapper[4767]: E0129 15:03:05.019434 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 29 15:03:06 crc kubenswrapper[4767]: I0129 15:03:06.018497 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:03:06 crc kubenswrapper[4767]: I0129 15:03:06.018506 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:06 crc kubenswrapper[4767]: I0129 15:03:06.020957 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 29 15:03:06 crc kubenswrapper[4767]: I0129 15:03:06.021008 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 29 15:03:06 crc kubenswrapper[4767]: I0129 15:03:06.021806 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 29 15:03:06 crc kubenswrapper[4767]: I0129 15:03:06.021982 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 29 15:03:06 crc kubenswrapper[4767]: I0129 15:03:06.032034 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:03:07 crc kubenswrapper[4767]: I0129 15:03:07.018502 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:07 crc kubenswrapper[4767]: I0129 15:03:07.018540 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:03:07 crc kubenswrapper[4767]: I0129 15:03:07.021428 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 29 15:03:07 crc kubenswrapper[4767]: I0129 15:03:07.021644 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 29 15:03:12 crc kubenswrapper[4767]: I0129 15:03:12.806253 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:03:12 crc kubenswrapper[4767]: I0129 15:03:12.806357 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.759251 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:13 crc kubenswrapper[4767]: E0129 15:03:13.759491 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:05:15.759457783 +0000 UTC m=+271.569774832 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.759761 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.768410 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.860394 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.860443 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.860473 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.862018 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.865450 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.866373 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:03:13 crc kubenswrapper[4767]: I0129 15:03:13.933888 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.143599 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.156320 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.312001 4767 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.365259 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.369791 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q9hsl"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.369976 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.370476 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-89fkh"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.371019 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.371593 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.379245 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.379807 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.380029 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.380169 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.380358 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.380571 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.380752 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.381143 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.381303 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.381436 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.381936 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.390957 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.391409 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.391546 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.391767 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.391924 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.391965 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.391986 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.391995 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.392618 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.392727 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.392885 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.396486 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.396639 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.397272 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.406759 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.406780 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.406835 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.411785 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vbh9z"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.412117 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.412562 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.413164 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.413409 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vbh9z" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.414434 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.414593 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.414963 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.415670 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.415912 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416328 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pvhqb"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416431 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgnbp\" (UniqueName: \"kubernetes.io/projected/861eefc8-26d1-4726-857d-1e44816cdadb-kube-api-access-pgnbp\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416476 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0aa203a3-ab0f-4174-b2dc-897b088b35cc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416498 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-trusted-ca-bundle\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416520 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-auth-proxy-config\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416537 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e3cbd296-0674-49b6-b9a5-7bb9f6736387-images\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416554 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llntk\" (UniqueName: \"kubernetes.io/projected/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-kube-api-access-llntk\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416574 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crwc9\" (UniqueName: \"kubernetes.io/projected/e3cbd296-0674-49b6-b9a5-7bb9f6736387-kube-api-access-crwc9\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416592 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6v4k\" (UniqueName: \"kubernetes.io/projected/0aa203a3-ab0f-4174-b2dc-897b088b35cc-kube-api-access-r6v4k\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416612 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-console-config\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416630 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-oauth-config\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416644 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-service-ca\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416658 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-oauth-serving-cert\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416676 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-serving-cert\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416693 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3cbd296-0674-49b6-b9a5-7bb9f6736387-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416715 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/315c0ef3-a03d-4195-bda6-d08219faeebf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416755 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0aa203a3-ab0f-4174-b2dc-897b088b35cc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416793 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgp96\" (UniqueName: \"kubernetes.io/projected/315c0ef3-a03d-4195-bda6-d08219faeebf-kube-api-access-kgp96\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416819 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0aa203a3-ab0f-4174-b2dc-897b088b35cc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416851 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cbd296-0674-49b6-b9a5-7bb9f6736387-config\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416800 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416870 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-config\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416925 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315c0ef3-a03d-4195-bda6-d08219faeebf-serving-cert\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416946 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-machine-approver-tls\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.416972 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.417296 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.418252 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.418410 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.418578 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.418698 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.419174 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.419346 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.419550 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.419867 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.422104 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.422362 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.423168 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.423517 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.423672 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.423843 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.423977 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.424083 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.424759 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9dkpb"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.425204 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.425424 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.425506 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.425674 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.425860 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.425431 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.426053 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.426193 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.426227 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-54tbh"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.426325 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.426494 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.426543 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.426626 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.427403 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.437739 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.438184 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4xbzf"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.438531 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.438614 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.440200 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.453788 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.454282 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.454413 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.454533 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.455212 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.455419 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.458288 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-5g8c7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.458749 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.462395 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qksvg"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.488359 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.490008 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.490467 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.491644 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.498985 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.499549 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.508495 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.510488 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.510617 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.511705 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.512460 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lxfft"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.531449 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.533404 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d348769a-75fd-4eab-abe7-2dee6bb14f73-config\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.537514 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.538041 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.538432 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.538788 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.539068 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.539203 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.540644 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.542608 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544235 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544385 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544467 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544573 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544585 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544645 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544720 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544750 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544886 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544612 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544500 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.544832 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.545194 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-config\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.545318 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.545384 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d348769a-75fd-4eab-abe7-2dee6bb14f73-trusted-ca\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.545407 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.545468 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-serving-cert\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.545487 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-client-ca\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.545505 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d348769a-75fd-4eab-abe7-2dee6bb14f73-serving-cert\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546097 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3cbd296-0674-49b6-b9a5-7bb9f6736387-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546126 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-config\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546244 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546275 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-policies\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546400 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ecb997-b521-4cda-a148-aa4b0f6caa72-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-njfmj\" (UID: \"b9ecb997-b521-4cda-a148-aa4b0f6caa72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546425 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8kn8\" (UniqueName: \"kubernetes.io/projected/e542b6c4-ef3e-4673-84ca-528e20243c1c-kube-api-access-p8kn8\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546443 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.546160 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.549413 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/315c0ef3-a03d-4195-bda6-d08219faeebf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.549484 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0aa203a3-ab0f-4174-b2dc-897b088b35cc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.549518 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e542b6c4-ef3e-4673-84ca-528e20243c1c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.549571 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58llh\" (UniqueName: \"kubernetes.io/projected/b9ecb997-b521-4cda-a148-aa4b0f6caa72-kube-api-access-58llh\") pod \"cluster-samples-operator-665b6dd947-njfmj\" (UID: \"b9ecb997-b521-4cda-a148-aa4b0f6caa72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.550538 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/315c0ef3-a03d-4195-bda6-d08219faeebf-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.550607 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tthx\" (UniqueName: \"kubernetes.io/projected/d8945cab-7833-45ef-b64d-597b5b0db3eb-kube-api-access-8tthx\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.550795 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.550907 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.551029 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0aa203a3-ab0f-4174-b2dc-897b088b35cc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.551123 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e542b6c4-ef3e-4673-84ca-528e20243c1c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.551218 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgp96\" (UniqueName: \"kubernetes.io/projected/315c0ef3-a03d-4195-bda6-d08219faeebf-kube-api-access-kgp96\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.551384 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cbd296-0674-49b6-b9a5-7bb9f6736387-config\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.551505 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-dir\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.551653 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-config\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.552449 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.552606 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-machine-approver-tls\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.552706 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.552813 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315c0ef3-a03d-4195-bda6-d08219faeebf-serving-cert\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.552951 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgnbp\" (UniqueName: \"kubernetes.io/projected/861eefc8-26d1-4726-857d-1e44816cdadb-kube-api-access-pgnbp\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553069 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-service-ca-bundle\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553176 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0aa203a3-ab0f-4174-b2dc-897b088b35cc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553270 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-trusted-ca-bundle\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553366 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e3cbd296-0674-49b6-b9a5-7bb9f6736387-images\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553463 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/393817f3-da00-4bd4-96ae-626b22da8e87-serving-cert\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553564 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18cf8157-44ad-4a44-8c51-9e41e1d50b52-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553667 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-auth-proxy-config\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553760 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.553872 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llntk\" (UniqueName: \"kubernetes.io/projected/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-kube-api-access-llntk\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.562646 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.562964 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18cf8157-44ad-4a44-8c51-9e41e1d50b52-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.563125 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crwc9\" (UniqueName: \"kubernetes.io/projected/e3cbd296-0674-49b6-b9a5-7bb9f6736387-kube-api-access-crwc9\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.557126 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-auth-proxy-config\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.559028 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0aa203a3-ab0f-4174-b2dc-897b088b35cc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.560030 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315c0ef3-a03d-4195-bda6-d08219faeebf-serving-cert\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.560084 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3cbd296-0674-49b6-b9a5-7bb9f6736387-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.560591 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3cbd296-0674-49b6-b9a5-7bb9f6736387-config\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561376 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-serving-cert\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561549 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561674 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-trusted-ca-bundle\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.562430 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-machine-approver-tls\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.555277 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0aa203a3-ab0f-4174-b2dc-897b088b35cc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.552395 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-config\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.556460 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e3cbd296-0674-49b6-b9a5-7bb9f6736387-images\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561556 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561602 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561599 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561641 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561662 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561670 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561688 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561699 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561711 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561778 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561790 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561829 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561837 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561913 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561962 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.561991 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.562017 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.562114 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.565064 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.569365 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.582789 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.587245 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.587987 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4rl7s"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.588400 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.588723 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.589764 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.591092 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.591498 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.591628 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.591760 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6v4k\" (UniqueName: \"kubernetes.io/projected/0aa203a3-ab0f-4174-b2dc-897b088b35cc-kube-api-access-r6v4k\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.591862 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-console-config\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.591982 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh8nj\" (UniqueName: \"kubernetes.io/projected/18cf8157-44ad-4a44-8c51-9e41e1d50b52-kube-api-access-jh8nj\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592076 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mcrc\" (UniqueName: \"kubernetes.io/projected/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-kube-api-access-2mcrc\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592173 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-oauth-config\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592268 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-service-ca\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592367 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhjvl\" (UniqueName: \"kubernetes.io/projected/737966ab-a636-4868-a7e9-6a2901383f73-kube-api-access-nhjvl\") pod \"downloads-7954f5f757-vbh9z\" (UID: \"737966ab-a636-4868-a7e9-6a2901383f73\") " pod="openshift-console/downloads-7954f5f757-vbh9z" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592445 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8945cab-7833-45ef-b64d-597b5b0db3eb-serving-cert\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592518 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592584 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-oauth-serving-cert\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592649 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v2cl\" (UniqueName: \"kubernetes.io/projected/393817f3-da00-4bd4-96ae-626b22da8e87-kube-api-access-7v2cl\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.592729 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsjjd\" (UniqueName: \"kubernetes.io/projected/d348769a-75fd-4eab-abe7-2dee6bb14f73-kube-api-access-lsjjd\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.593939 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.594737 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-service-ca\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.595082 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-oauth-serving-cert\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.595614 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-console-config\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.601155 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-89fkh"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.603455 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.604282 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.598145 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.601571 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-oauth-config\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.608006 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fkbz7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.608819 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.614006 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.614379 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.614591 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.614929 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.617491 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.634269 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.638281 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.643357 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.643809 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.646851 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q9hsl"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.648242 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.649396 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.650193 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.650591 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.650948 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.651510 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.651882 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.652429 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.652499 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.654260 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jm5c7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.654709 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.654826 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tsjd8"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.655336 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.655375 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.655991 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.656368 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.657343 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rmw5d"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.657949 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.658265 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.658801 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.661560 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pvhqb"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.666192 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.669031 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.670866 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.673526 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.675687 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.676942 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.678042 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.679202 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-54tbh"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.680476 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.681444 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vbh9z"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.682668 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9dkpb"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.684143 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.684962 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.686098 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4xbzf"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.687697 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.689376 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qksvg"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.690715 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bk5s7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.692398 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.692492 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.692516 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.695313 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kk65s"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.695797 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.695825 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.698823 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.699404 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.701183 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.703002 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lxfft"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.703813 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.704406 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpww7\" (UniqueName: \"kubernetes.io/projected/2356f459-9a7b-4d07-ba43-135788ee3c91-kube-api-access-dpww7\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.704445 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsjjd\" (UniqueName: \"kubernetes.io/projected/d348769a-75fd-4eab-abe7-2dee6bb14f73-kube-api-access-lsjjd\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.704841 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.705764 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.705828 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94r67\" (UniqueName: \"kubernetes.io/projected/53138b80-9329-4c54-b607-3df3a694688f-kube-api-access-94r67\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.705870 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c481090-cd6e-428d-bd33-56a4fb452535-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5vkq5\" (UID: \"0c481090-cd6e-428d-bd33-56a4fb452535\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706009 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2356f459-9a7b-4d07-ba43-135788ee3c91-node-pullsecrets\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706110 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53138b80-9329-4c54-b607-3df3a694688f-proxy-tls\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706142 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw794\" (UniqueName: \"kubernetes.io/projected/bd5365ec-5c61-4e64-b3de-d6fd2498084e-kube-api-access-fw794\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706208 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/97277fd3-32c2-421d-8a22-99d25e91e14a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706270 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-serving-cert\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706337 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706423 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91a22e1c-1a0b-46ac-a8ad-54918a45671d-service-ca-bundle\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706583 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd5365ec-5c61-4e64-b3de-d6fd2498084e-signing-key\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706632 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-client\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706687 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d348769a-75fd-4eab-abe7-2dee6bb14f73-serving-cert\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.706735 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-config\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.707383 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-config\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.707444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.707504 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x5ft\" (UniqueName: \"kubernetes.io/projected/e9e8e987-a20f-4f6a-8102-7782d84f88e7-kube-api-access-5x5ft\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.707531 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-encryption-config\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.707597 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ecb997-b521-4cda-a148-aa4b0f6caa72-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-njfmj\" (UID: \"b9ecb997-b521-4cda-a148-aa4b0f6caa72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708299 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kk65s"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708319 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53138b80-9329-4c54-b607-3df3a694688f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708354 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrp9m\" (UniqueName: \"kubernetes.io/projected/718554cf-a4a8-4a16-b070-64fb54973298-kube-api-access-rrp9m\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708380 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8kn8\" (UniqueName: \"kubernetes.io/projected/e542b6c4-ef3e-4673-84ca-528e20243c1c-kube-api-access-p8kn8\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708399 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v8gg\" (UniqueName: \"kubernetes.io/projected/97277fd3-32c2-421d-8a22-99d25e91e14a-kube-api-access-5v8gg\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708430 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55l4m\" (UniqueName: \"kubernetes.io/projected/61181831-03f4-4031-8003-0767b3a18856-kube-api-access-55l4m\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708564 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7885\" (UniqueName: \"kubernetes.io/projected/1263384a-6a0a-4529-ad99-3162b86e0c20-kube-api-access-m7885\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708612 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2tfb\" (UniqueName: \"kubernetes.io/projected/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-kube-api-access-f2tfb\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708634 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tthx\" (UniqueName: \"kubernetes.io/projected/d8945cab-7833-45ef-b64d-597b5b0db3eb-kube-api-access-8tthx\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708656 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e542b6c4-ef3e-4673-84ca-528e20243c1c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708699 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-dir\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708730 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-service-ca-bundle\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708750 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1263384a-6a0a-4529-ad99-3162b86e0c20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708770 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-service-ca\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708789 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/393817f3-da00-4bd4-96ae-626b22da8e87-serving-cert\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708807 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18cf8157-44ad-4a44-8c51-9e41e1d50b52-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708830 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708861 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708877 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-ca\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708914 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-audit-policies\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708934 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708962 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh8nj\" (UniqueName: \"kubernetes.io/projected/18cf8157-44ad-4a44-8c51-9e41e1d50b52-kube-api-access-jh8nj\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708979 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.708996 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.709016 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8945cab-7833-45ef-b64d-597b5b0db3eb-serving-cert\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.709036 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-etcd-serving-ca\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.709063 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v2cl\" (UniqueName: \"kubernetes.io/projected/393817f3-da00-4bd4-96ae-626b22da8e87-kube-api-access-7v2cl\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.710916 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/b9ecb997-b521-4cda-a148-aa4b0f6caa72-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-njfmj\" (UID: \"b9ecb997-b521-4cda-a148-aa4b0f6caa72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.710939 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d348769a-75fd-4eab-abe7-2dee6bb14f73-serving-cert\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.711027 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-dir\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.711217 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/393817f3-da00-4bd4-96ae-626b22da8e87-service-ca-bundle\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.711355 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.711930 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18cf8157-44ad-4a44-8c51-9e41e1d50b52-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.714647 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.714844 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8945cab-7833-45ef-b64d-597b5b0db3eb-serving-cert\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.716367 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.716596 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.716717 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/393817f3-da00-4bd4-96ae-626b22da8e87-serving-cert\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719572 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc724\" (UniqueName: \"kubernetes.io/projected/4199f360-d546-466f-9dc2-dbd7c0756d93-kube-api-access-qc724\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719688 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58bph\" (UniqueName: \"kubernetes.io/projected/8c6ee408-b89d-4edf-b18f-d139046e8ccf-kube-api-access-58bph\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719748 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d348769a-75fd-4eab-abe7-2dee6bb14f73-config\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719784 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-config\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719837 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-config\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719946 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719987 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-audit\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720015 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-client-ca\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720048 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f7932acc-0464-45d2-886a-aefd898e00d0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lxfft\" (UID: \"f7932acc-0464-45d2-886a-aefd898e00d0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720086 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d348769a-75fd-4eab-abe7-2dee6bb14f73-trusted-ca\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720119 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720163 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-client-ca\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720202 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2356f459-9a7b-4d07-ba43-135788ee3c91-audit-dir\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720245 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/97277fd3-32c2-421d-8a22-99d25e91e14a-srv-cert\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720284 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-policies\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720313 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/718554cf-a4a8-4a16-b070-64fb54973298-webhook-cert\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.719866 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720360 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6vf8\" (UniqueName: \"kubernetes.io/projected/f7932acc-0464-45d2-886a-aefd898e00d0-kube-api-access-r6vf8\") pod \"multus-admission-controller-857f4d67dd-lxfft\" (UID: \"f7932acc-0464-45d2-886a-aefd898e00d0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720411 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720463 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9e8e987-a20f-4f6a-8102-7782d84f88e7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720490 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd5365ec-5c61-4e64-b3de-d6fd2498084e-signing-cabundle\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720514 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61181831-03f4-4031-8003-0767b3a18856-audit-dir\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720528 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d348769a-75fd-4eab-abe7-2dee6bb14f73-config\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720563 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e542b6c4-ef3e-4673-84ca-528e20243c1c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720591 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-etcd-client\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720611 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx7z6\" (UniqueName: \"kubernetes.io/projected/0c481090-cd6e-428d-bd33-56a4fb452535-kube-api-access-nx7z6\") pod \"package-server-manager-789f6589d5-5vkq5\" (UID: \"0c481090-cd6e-428d-bd33-56a4fb452535\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720633 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5n27\" (UniqueName: \"kubernetes.io/projected/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-kube-api-access-x5n27\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720659 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58llh\" (UniqueName: \"kubernetes.io/projected/b9ecb997-b521-4cda-a148-aa4b0f6caa72-kube-api-access-58llh\") pod \"cluster-samples-operator-665b6dd947-njfmj\" (UID: \"b9ecb997-b521-4cda-a148-aa4b0f6caa72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720681 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-metrics-certs\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720709 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4199f360-d546-466f-9dc2-dbd7c0756d93-serving-cert\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720726 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c6ee408-b89d-4edf-b18f-d139046e8ccf-secret-volume\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720747 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720770 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720791 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/718554cf-a4a8-4a16-b070-64fb54973298-apiservice-cert\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720827 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/718554cf-a4a8-4a16-b070-64fb54973298-tmpfs\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720869 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-serving-cert\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720911 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-encryption-config\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720936 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c6ee408-b89d-4edf-b18f-d139046e8ccf-config-volume\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720958 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-config\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.720986 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721005 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721026 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-image-import-ca\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721066 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721093 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9e8e987-a20f-4f6a-8102-7782d84f88e7-metrics-tls\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721110 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-serving-cert\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721135 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/53138b80-9329-4c54-b607-3df3a694688f-images\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721171 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtmbl\" (UniqueName: \"kubernetes.io/projected/91a22e1c-1a0b-46ac-a8ad-54918a45671d-kube-api-access-qtmbl\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721190 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1263384a-6a0a-4529-ad99-3162b86e0c20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721252 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9e8e987-a20f-4f6a-8102-7782d84f88e7-trusted-ca\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721278 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18cf8157-44ad-4a44-8c51-9e41e1d50b52-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721297 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-stats-auth\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721323 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcw8v\" (UniqueName: \"kubernetes.io/projected/1e502829-2ef5-4515-b599-e455100d1e38-kube-api-access-rcw8v\") pod \"migrator-59844c95c7-h26ds\" (UID: \"1e502829-2ef5-4515-b599-e455100d1e38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721344 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-config\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721371 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-default-certificate\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721393 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-serving-cert\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721415 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-config\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721435 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-etcd-client\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721456 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721466 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-config\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721483 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhjvl\" (UniqueName: \"kubernetes.io/projected/737966ab-a636-4868-a7e9-6a2901383f73-kube-api-access-nhjvl\") pod \"downloads-7954f5f757-vbh9z\" (UID: \"737966ab-a636-4868-a7e9-6a2901383f73\") " pod="openshift-console/downloads-7954f5f757-vbh9z" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721496 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721660 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.721506 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mcrc\" (UniqueName: \"kubernetes.io/projected/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-kube-api-access-2mcrc\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.722447 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e542b6c4-ef3e-4673-84ca-528e20243c1c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.722490 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tsjd8"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.722694 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e542b6c4-ef3e-4673-84ca-528e20243c1c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.723217 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4rl7s"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.723945 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.723975 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.724622 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-client-ca\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.724764 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-policies\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.724839 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.725679 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d348769a-75fd-4eab-abe7-2dee6bb14f73-trusted-ca\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.726103 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18cf8157-44ad-4a44-8c51-9e41e1d50b52-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.726763 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fkbz7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.729349 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.729422 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.729531 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.730835 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.732504 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bk5s7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.732788 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.733665 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.737134 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-5qnld"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.738124 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.738556 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-qq48g"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.739123 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.739838 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jm5c7"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.741000 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rmw5d"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.742020 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5qnld"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.743019 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg"] Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.753504 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.773712 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.793740 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.813742 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823197 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-metrics-certs\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823240 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b68b8\" (UniqueName: \"kubernetes.io/projected/e695aba7-f235-4206-9e18-c4cbc877d885-kube-api-access-b68b8\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823269 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4199f360-d546-466f-9dc2-dbd7c0756d93-serving-cert\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823295 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c6ee408-b89d-4edf-b18f-d139046e8ccf-secret-volume\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823320 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3affdc50-c158-41f5-a067-c0e3a61541d5-config-volume\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823485 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/718554cf-a4a8-4a16-b070-64fb54973298-apiservice-cert\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823611 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/718554cf-a4a8-4a16-b070-64fb54973298-tmpfs\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823663 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-serving-cert\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823684 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-config\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823708 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp929\" (UniqueName: \"kubernetes.io/projected/69d2b93c-da14-4c39-80a2-095dae5a1a42-kube-api-access-lp929\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823731 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-encryption-config\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823751 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c6ee408-b89d-4edf-b18f-d139046e8ccf-config-volume\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823775 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e695aba7-f235-4206-9e18-c4cbc877d885-profile-collector-cert\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823802 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823826 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-image-import-ca\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823847 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68fnf\" (UniqueName: \"kubernetes.io/projected/3e628934-9e7a-4c1e-b9fa-018c806f75da-kube-api-access-68fnf\") pod \"dns-operator-744455d44c-jm5c7\" (UID: \"3e628934-9e7a-4c1e-b9fa-018c806f75da\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823874 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-registration-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823916 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/53138b80-9329-4c54-b607-3df3a694688f-images\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823941 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d815d7e-754d-454d-8b2b-5891d1fef162-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.823963 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3e628934-9e7a-4c1e-b9fa-018c806f75da-metrics-tls\") pod \"dns-operator-744455d44c-jm5c7\" (UID: \"3e628934-9e7a-4c1e-b9fa-018c806f75da\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824020 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9e8e987-a20f-4f6a-8102-7782d84f88e7-metrics-tls\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824045 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-serving-cert\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824069 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-node-bootstrap-token\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824096 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtmbl\" (UniqueName: \"kubernetes.io/projected/91a22e1c-1a0b-46ac-a8ad-54918a45671d-kube-api-access-qtmbl\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824121 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05e72201-53f8-42f8-8a74-016cb6506c9f-config\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824140 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/718554cf-a4a8-4a16-b070-64fb54973298-tmpfs\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824146 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9e8e987-a20f-4f6a-8102-7782d84f88e7-trusted-ca\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824196 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1263384a-6a0a-4529-ad99-3162b86e0c20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824244 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-stats-auth\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824267 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcw8v\" (UniqueName: \"kubernetes.io/projected/1e502829-2ef5-4515-b599-e455100d1e38-kube-api-access-rcw8v\") pod \"migrator-59844c95c7-h26ds\" (UID: \"1e502829-2ef5-4515-b599-e455100d1e38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824289 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05e72201-53f8-42f8-8a74-016cb6506c9f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824309 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg2sw\" (UniqueName: \"kubernetes.io/projected/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-kube-api-access-pg2sw\") pod \"ingress-canary-kk65s\" (UID: \"183bee6d-c673-4a5a-a0da-1d3bc44ead6b\") " pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824327 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b8a5533d-d711-4fa8-ab7a-a686f77cb578-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qxdzw\" (UID: \"b8a5533d-d711-4fa8-ab7a-a686f77cb578\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824345 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-config\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824367 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-default-certificate\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824383 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-serving-cert\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824403 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-etcd-client\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824419 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824437 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-config\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824480 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpww7\" (UniqueName: \"kubernetes.io/projected/2356f459-9a7b-4d07-ba43-135788ee3c91-kube-api-access-dpww7\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824520 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94r67\" (UniqueName: \"kubernetes.io/projected/53138b80-9329-4c54-b607-3df3a694688f-kube-api-access-94r67\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824564 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c481090-cd6e-428d-bd33-56a4fb452535-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5vkq5\" (UID: \"0c481090-cd6e-428d-bd33-56a4fb452535\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824597 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4pvs\" (UniqueName: \"kubernetes.io/projected/b8a5533d-d711-4fa8-ab7a-a686f77cb578-kube-api-access-f4pvs\") pod \"control-plane-machine-set-operator-78cbb6b69f-qxdzw\" (UID: \"b8a5533d-d711-4fa8-ab7a-a686f77cb578\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824627 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2356f459-9a7b-4d07-ba43-135788ee3c91-node-pullsecrets\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824655 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53138b80-9329-4c54-b607-3df3a694688f-proxy-tls\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824679 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw794\" (UniqueName: \"kubernetes.io/projected/bd5365ec-5c61-4e64-b3de-d6fd2498084e-kube-api-access-fw794\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824705 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/97277fd3-32c2-421d-8a22-99d25e91e14a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824738 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-serving-cert\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824766 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91a22e1c-1a0b-46ac-a8ad-54918a45671d-service-ca-bundle\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824792 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd5365ec-5c61-4e64-b3de-d6fd2498084e-signing-key\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824815 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-client\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824877 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r52x\" (UniqueName: \"kubernetes.io/projected/3affdc50-c158-41f5-a067-c0e3a61541d5-kube-api-access-8r52x\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824937 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x5ft\" (UniqueName: \"kubernetes.io/projected/e9e8e987-a20f-4f6a-8102-7782d84f88e7-kube-api-access-5x5ft\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.824981 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-image-import-ca\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825130 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2356f459-9a7b-4d07-ba43-135788ee3c91-node-pullsecrets\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825158 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-encryption-config\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825179 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3affdc50-c158-41f5-a067-c0e3a61541d5-metrics-tls\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825216 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53138b80-9329-4c54-b607-3df3a694688f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825242 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrp9m\" (UniqueName: \"kubernetes.io/projected/718554cf-a4a8-4a16-b070-64fb54973298-kube-api-access-rrp9m\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825275 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v8gg\" (UniqueName: \"kubernetes.io/projected/97277fd3-32c2-421d-8a22-99d25e91e14a-kube-api-access-5v8gg\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825300 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2tfb\" (UniqueName: \"kubernetes.io/projected/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-kube-api-access-f2tfb\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825315 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e9e8e987-a20f-4f6a-8102-7782d84f88e7-trusted-ca\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825375 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhj9r\" (UniqueName: \"kubernetes.io/projected/ed733af7-90b8-481c-86ea-e910a53c36de-kube-api-access-bhj9r\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825398 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55l4m\" (UniqueName: \"kubernetes.io/projected/61181831-03f4-4031-8003-0767b3a18856-kube-api-access-55l4m\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825416 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7885\" (UniqueName: \"kubernetes.io/projected/1263384a-6a0a-4529-ad99-3162b86e0c20-kube-api-access-m7885\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.825997 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91a22e1c-1a0b-46ac-a8ad-54918a45671d-service-ca-bundle\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.827007 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-config\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.827302 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.827873 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-serving-cert\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828030 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-encryption-config\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828253 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1263384a-6a0a-4529-ad99-3162b86e0c20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828289 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-service-ca\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828358 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-socket-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828418 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d815d7e-754d-454d-8b2b-5891d1fef162-proxy-tls\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828438 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-mountpoint-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828484 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-ca\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828509 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-audit-policies\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828726 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-certs\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828761 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-csi-data-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828789 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-cert\") pod \"ingress-canary-kk65s\" (UID: \"183bee6d-c673-4a5a-a0da-1d3bc44ead6b\") " pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828833 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828869 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-etcd-serving-ca\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828922 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05e72201-53f8-42f8-8a74-016cb6506c9f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828962 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc724\" (UniqueName: \"kubernetes.io/projected/4199f360-d546-466f-9dc2-dbd7c0756d93-kube-api-access-qc724\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828962 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4199f360-d546-466f-9dc2-dbd7c0756d93-serving-cert\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828832 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-etcd-client\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.828999 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-metrics-certs\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829031 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58bph\" (UniqueName: \"kubernetes.io/projected/8c6ee408-b89d-4edf-b18f-d139046e8ccf-kube-api-access-58bph\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829052 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e643b7f-8f80-4aed-a61f-13ffa230b885-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829080 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-config\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829099 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-audit\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829116 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e643b7f-8f80-4aed-a61f-13ffa230b885-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829144 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f7932acc-0464-45d2-886a-aefd898e00d0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lxfft\" (UID: \"f7932acc-0464-45d2-886a-aefd898e00d0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829162 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e643b7f-8f80-4aed-a61f-13ffa230b885-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829284 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-client-ca\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829320 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2356f459-9a7b-4d07-ba43-135788ee3c91-audit-dir\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829343 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hglpd\" (UniqueName: \"kubernetes.io/projected/3d815d7e-754d-454d-8b2b-5891d1fef162-kube-api-access-hglpd\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829362 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/718554cf-a4a8-4a16-b070-64fb54973298-webhook-cert\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829379 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/97277fd3-32c2-421d-8a22-99d25e91e14a-srv-cert\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829396 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6vf8\" (UniqueName: \"kubernetes.io/projected/f7932acc-0464-45d2-886a-aefd898e00d0-kube-api-access-r6vf8\") pod \"multus-admission-controller-857f4d67dd-lxfft\" (UID: \"f7932acc-0464-45d2-886a-aefd898e00d0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.829792 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.830379 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-serving-cert\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.830481 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-audit\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.830993 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-audit-policies\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831160 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-default-certificate\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831349 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9e8e987-a20f-4f6a-8102-7782d84f88e7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831379 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61181831-03f4-4031-8003-0767b3a18856-audit-dir\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831399 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd5365ec-5c61-4e64-b3de-d6fd2498084e-signing-cabundle\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831448 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-plugins-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831470 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx7z6\" (UniqueName: \"kubernetes.io/projected/0c481090-cd6e-428d-bd33-56a4fb452535-kube-api-access-nx7z6\") pod \"package-server-manager-789f6589d5-5vkq5\" (UID: \"0c481090-cd6e-428d-bd33-56a4fb452535\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831490 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5n27\" (UniqueName: \"kubernetes.io/projected/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-kube-api-access-x5n27\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831489 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e9e8e987-a20f-4f6a-8102-7782d84f88e7-metrics-tls\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831508 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e695aba7-f235-4206-9e18-c4cbc877d885-srv-cert\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831523 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2356f459-9a7b-4d07-ba43-135788ee3c91-audit-dir\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831576 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-etcd-client\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831642 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/61181831-03f4-4031-8003-0767b3a18856-audit-dir\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831928 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-client-ca\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.831947 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2356f459-9a7b-4d07-ba43-135788ee3c91-etcd-serving-ca\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.832531 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/61181831-03f4-4031-8003-0767b3a18856-encryption-config\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.833281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/91a22e1c-1a0b-46ac-a8ad-54918a45671d-stats-auth\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.833659 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f7932acc-0464-45d2-886a-aefd898e00d0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lxfft\" (UID: \"f7932acc-0464-45d2-886a-aefd898e00d0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.833946 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.833976 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2356f459-9a7b-4d07-ba43-135788ee3c91-etcd-client\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.835414 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61181831-03f4-4031-8003-0767b3a18856-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.835631 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/53138b80-9329-4c54-b607-3df3a694688f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.836283 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-config\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.853344 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.855418 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/53138b80-9329-4c54-b607-3df3a694688f-images\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.872934 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.894106 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.897577 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1263384a-6a0a-4529-ad99-3162b86e0c20-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.908700 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"59ceb084d6d6aab93313e60b322dbb6f0a22630002453a7e4399651af2c98509"} Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.909495 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"dcc3bc9dd3a2607a5e6dbd6f4240958d6b7999cf871aa79297cf80dbf9b97239"} Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.910833 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"70be4d3bed584a388597fd78648bb190a0f4d3a83111003f0d06c4ca496eff09"} Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.910860 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a494f829921219b6619c558857598378caa947181bd064795a950c9bbd1a1c7d"} Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.911096 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.913148 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.932829 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e695aba7-f235-4206-9e18-c4cbc877d885-profile-collector-cert\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.932878 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68fnf\" (UniqueName: \"kubernetes.io/projected/3e628934-9e7a-4c1e-b9fa-018c806f75da-kube-api-access-68fnf\") pod \"dns-operator-744455d44c-jm5c7\" (UID: \"3e628934-9e7a-4c1e-b9fa-018c806f75da\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933061 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933094 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-registration-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933175 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d815d7e-754d-454d-8b2b-5891d1fef162-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933196 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3e628934-9e7a-4c1e-b9fa-018c806f75da-metrics-tls\") pod \"dns-operator-744455d44c-jm5c7\" (UID: \"3e628934-9e7a-4c1e-b9fa-018c806f75da\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933217 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05e72201-53f8-42f8-8a74-016cb6506c9f-config\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933234 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-node-bootstrap-token\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933264 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b8a5533d-d711-4fa8-ab7a-a686f77cb578-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qxdzw\" (UID: \"b8a5533d-d711-4fa8-ab7a-a686f77cb578\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933373 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05e72201-53f8-42f8-8a74-016cb6506c9f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933394 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg2sw\" (UniqueName: \"kubernetes.io/projected/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-kube-api-access-pg2sw\") pod \"ingress-canary-kk65s\" (UID: \"183bee6d-c673-4a5a-a0da-1d3bc44ead6b\") " pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933507 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4pvs\" (UniqueName: \"kubernetes.io/projected/b8a5533d-d711-4fa8-ab7a-a686f77cb578-kube-api-access-f4pvs\") pod \"control-plane-machine-set-operator-78cbb6b69f-qxdzw\" (UID: \"b8a5533d-d711-4fa8-ab7a-a686f77cb578\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933560 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3affdc50-c158-41f5-a067-c0e3a61541d5-metrics-tls\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933578 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r52x\" (UniqueName: \"kubernetes.io/projected/3affdc50-c158-41f5-a067-c0e3a61541d5-kube-api-access-8r52x\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933638 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhj9r\" (UniqueName: \"kubernetes.io/projected/ed733af7-90b8-481c-86ea-e910a53c36de-kube-api-access-bhj9r\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933713 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-socket-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933757 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d815d7e-754d-454d-8b2b-5891d1fef162-proxy-tls\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933776 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-mountpoint-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933866 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-certs\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933883 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-cert\") pod \"ingress-canary-kk65s\" (UID: \"183bee6d-c673-4a5a-a0da-1d3bc44ead6b\") " pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933911 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-socket-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933941 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-mountpoint-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.933964 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-csi-data-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934017 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-csi-data-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934039 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05e72201-53f8-42f8-8a74-016cb6506c9f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934056 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-registration-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934069 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e643b7f-8f80-4aed-a61f-13ffa230b885-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934115 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e643b7f-8f80-4aed-a61f-13ffa230b885-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934142 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e643b7f-8f80-4aed-a61f-13ffa230b885-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934174 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hglpd\" (UniqueName: \"kubernetes.io/projected/3d815d7e-754d-454d-8b2b-5891d1fef162-kube-api-access-hglpd\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934226 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-plugins-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934260 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e695aba7-f235-4206-9e18-c4cbc877d885-srv-cert\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934290 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b68b8\" (UniqueName: \"kubernetes.io/projected/e695aba7-f235-4206-9e18-c4cbc877d885-kube-api-access-b68b8\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934305 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ed733af7-90b8-481c-86ea-e910a53c36de-plugins-dir\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934322 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3affdc50-c158-41f5-a067-c0e3a61541d5-config-volume\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934352 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d815d7e-754d-454d-8b2b-5891d1fef162-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.934379 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp929\" (UniqueName: \"kubernetes.io/projected/69d2b93c-da14-4c39-80a2-095dae5a1a42-kube-api-access-lp929\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.942496 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1263384a-6a0a-4529-ad99-3162b86e0c20-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.954837 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.973044 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.986008 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/97277fd3-32c2-421d-8a22-99d25e91e14a-srv-cert\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.993610 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.998066 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c6ee408-b89d-4edf-b18f-d139046e8ccf-secret-volume\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:14 crc kubenswrapper[4767]: I0129 15:03:14.998211 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e695aba7-f235-4206-9e18-c4cbc877d885-profile-collector-cert\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.002450 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/97277fd3-32c2-421d-8a22-99d25e91e14a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.014638 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.035372 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.053021 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.063024 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/53138b80-9329-4c54-b607-3df3a694688f-proxy-tls\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.086765 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0aa203a3-ab0f-4174-b2dc-897b088b35cc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.107318 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgp96\" (UniqueName: \"kubernetes.io/projected/315c0ef3-a03d-4195-bda6-d08219faeebf-kube-api-access-kgp96\") pod \"openshift-config-operator-7777fb866f-ssmkq\" (UID: \"315c0ef3-a03d-4195-bda6-d08219faeebf\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.128513 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgnbp\" (UniqueName: \"kubernetes.io/projected/861eefc8-26d1-4726-857d-1e44816cdadb-kube-api-access-pgnbp\") pod \"console-f9d7485db-89fkh\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.154082 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llntk\" (UniqueName: \"kubernetes.io/projected/3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2-kube-api-access-llntk\") pod \"machine-approver-56656f9798-wfdcw\" (UID: \"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.189942 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crwc9\" (UniqueName: \"kubernetes.io/projected/e3cbd296-0674-49b6-b9a5-7bb9f6736387-kube-api-access-crwc9\") pod \"machine-api-operator-5694c8668f-q9hsl\" (UID: \"e3cbd296-0674-49b6-b9a5-7bb9f6736387\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.189971 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.193205 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.214227 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.222175 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd5365ec-5c61-4e64-b3de-d6fd2498084e-signing-key\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.233681 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.248096 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd5365ec-5c61-4e64-b3de-d6fd2498084e-signing-cabundle\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.253668 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.274043 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.293287 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.309402 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0c481090-cd6e-428d-bd33-56a4fb452535-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-5vkq5\" (UID: \"0c481090-cd6e-428d-bd33-56a4fb452535\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.324809 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.335239 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6v4k\" (UniqueName: \"kubernetes.io/projected/0aa203a3-ab0f-4174-b2dc-897b088b35cc-kube-api-access-r6v4k\") pod \"cluster-image-registry-operator-dc59b4c8b-dmgtc\" (UID: \"0aa203a3-ab0f-4174-b2dc-897b088b35cc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.335545 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.338289 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq"] Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.353590 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.359280 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.374057 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.382728 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-serving-cert\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.393162 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.394595 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-config\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.413178 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.433190 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.441668 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-ca\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.441885 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.453070 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.465189 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.474936 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.483461 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-serving-cert\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.493945 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.506209 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-client\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.517198 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.522661 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-etcd-service-ca\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.538725 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.553597 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.574204 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.578307 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-89fkh"] Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.581386 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-config\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.593653 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.613683 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.619932 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c6ee408-b89d-4edf-b18f-d139046e8ccf-config-volume\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.632031 4767 request.go:700] Waited for 1.003858511s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpackageserver-service-cert&limit=500&resourceVersion=0 Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.634399 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.634682 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc"] Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.645772 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/718554cf-a4a8-4a16-b070-64fb54973298-webhook-cert\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.647336 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/718554cf-a4a8-4a16-b070-64fb54973298-apiservice-cert\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.656602 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.672046 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q9hsl"] Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.673787 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 29 15:03:15 crc kubenswrapper[4767]: W0129 15:03:15.691384 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3cbd296_0674_49b6_b9a5_7bb9f6736387.slice/crio-e62b52c42e011a9b60a039b267c38c7c44afbc8fa0a7d5033b1b3cb7a885ba49 WatchSource:0}: Error finding container e62b52c42e011a9b60a039b267c38c7c44afbc8fa0a7d5033b1b3cb7a885ba49: Status 404 returned error can't find the container with id e62b52c42e011a9b60a039b267c38c7c44afbc8fa0a7d5033b1b3cb7a885ba49 Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.693424 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.713634 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.719169 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e695aba7-f235-4206-9e18-c4cbc877d885-srv-cert\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.734194 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.754636 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.773356 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d815d7e-754d-454d-8b2b-5891d1fef162-proxy-tls\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.773533 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.797455 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.813420 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.818492 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05e72201-53f8-42f8-8a74-016cb6506c9f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.832713 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.833887 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05e72201-53f8-42f8-8a74-016cb6506c9f-config\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.854087 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.855418 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e643b7f-8f80-4aed-a61f-13ffa230b885-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.875570 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.895539 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.909652 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8e643b7f-8f80-4aed-a61f-13ffa230b885-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.914347 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.925490 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" event={"ID":"0aa203a3-ab0f-4174-b2dc-897b088b35cc","Type":"ContainerStarted","Data":"064ba466fc2160fdbac2bc25944596d0180210ebc5cd7131b6e35836d9769a38"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.925585 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" event={"ID":"0aa203a3-ab0f-4174-b2dc-897b088b35cc","Type":"ContainerStarted","Data":"90a1dae0d5229bb27ad608d6d25f9694da882598d0b534cdea4fa78c1476f392"} Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.933941 4767 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934028 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-node-bootstrap-token podName:69d2b93c-da14-4c39-80a2-095dae5a1a42 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:16.434008907 +0000 UTC m=+152.244325946 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-node-bootstrap-token") pod "machine-config-server-qq48g" (UID: "69d2b93c-da14-4c39-80a2-095dae5a1a42") : failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.934383 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934401 4767 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934435 4767 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934439 4767 secret.go:188] Couldn't get secret openshift-dns-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934416 4767 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934502 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b8a5533d-d711-4fa8-ab7a-a686f77cb578-control-plane-machine-set-operator-tls podName:b8a5533d-d711-4fa8-ab7a-a686f77cb578 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:16.43447233 +0000 UTC m=+152.244789369 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/b8a5533d-d711-4fa8-ab7a-a686f77cb578-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-qxdzw" (UID: "b8a5533d-d711-4fa8-ab7a-a686f77cb578") : failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934527 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-certs podName:69d2b93c-da14-4c39-80a2-095dae5a1a42 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:16.434513801 +0000 UTC m=+152.244830850 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-certs") pod "machine-config-server-qq48g" (UID: "69d2b93c-da14-4c39-80a2-095dae5a1a42") : failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934548 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3affdc50-c158-41f5-a067-c0e3a61541d5-metrics-tls podName:3affdc50-c158-41f5-a067-c0e3a61541d5 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:16.434540092 +0000 UTC m=+152.244857141 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/3affdc50-c158-41f5-a067-c0e3a61541d5-metrics-tls") pod "dns-default-5qnld" (UID: "3affdc50-c158-41f5-a067-c0e3a61541d5") : failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934554 4767 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934565 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e628934-9e7a-4c1e-b9fa-018c806f75da-metrics-tls podName:3e628934-9e7a-4c1e-b9fa-018c806f75da nodeName:}" failed. No retries permitted until 2026-01-29 15:03:16.434556722 +0000 UTC m=+152.244873771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/3e628934-9e7a-4c1e-b9fa-018c806f75da-metrics-tls") pod "dns-operator-744455d44c-jm5c7" (UID: "3e628934-9e7a-4c1e-b9fa-018c806f75da") : failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.934691 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3affdc50-c158-41f5-a067-c0e3a61541d5-config-volume podName:3affdc50-c158-41f5-a067-c0e3a61541d5 nodeName:}" failed. No retries permitted until 2026-01-29 15:03:16.434669656 +0000 UTC m=+152.244986705 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/3affdc50-c158-41f5-a067-c0e3a61541d5-config-volume") pod "dns-default-5qnld" (UID: "3affdc50-c158-41f5-a067-c0e3a61541d5") : failed to sync configmap cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.934969 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-89fkh" event={"ID":"861eefc8-26d1-4726-857d-1e44816cdadb","Type":"ContainerStarted","Data":"cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.935039 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-89fkh" event={"ID":"861eefc8-26d1-4726-857d-1e44816cdadb","Type":"ContainerStarted","Data":"48594a6f53da007ea41a8de4b1e1669fec04810c6b1705cd7f5bbfef0465ff3e"} Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.935245 4767 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: E0129 15:03:15.935310 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-cert podName:183bee6d-c673-4a5a-a0da-1d3bc44ead6b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:16.435293564 +0000 UTC m=+152.245610623 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-cert") pod "ingress-canary-kk65s" (UID: "183bee6d-c673-4a5a-a0da-1d3bc44ead6b") : failed to sync secret cache: timed out waiting for the condition Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.938965 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" event={"ID":"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2","Type":"ContainerStarted","Data":"e56d55894c09d9e7efe0836bac60b20d952691c54c24d9f3529e0e6f22b0b93d"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.939035 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" event={"ID":"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2","Type":"ContainerStarted","Data":"47e90232891e3241ef8f5d0cef406067b1b5d3cd22aeb6e521b374bbbe498e11"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.943959 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" event={"ID":"e3cbd296-0674-49b6-b9a5-7bb9f6736387","Type":"ContainerStarted","Data":"fd5ac49410dcf2d844d89e43aa97eb8a3fe09484b8f93279209027da9eb8207c"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.944048 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" event={"ID":"e3cbd296-0674-49b6-b9a5-7bb9f6736387","Type":"ContainerStarted","Data":"e62b52c42e011a9b60a039b267c38c7c44afbc8fa0a7d5033b1b3cb7a885ba49"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.948865 4767 generic.go:334] "Generic (PLEG): container finished" podID="315c0ef3-a03d-4195-bda6-d08219faeebf" containerID="9b619fa695f344dc2cc2d240c727248a175ff67ae7507e35b8d03ee20130a627" exitCode=0 Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.949165 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" event={"ID":"315c0ef3-a03d-4195-bda6-d08219faeebf","Type":"ContainerDied","Data":"9b619fa695f344dc2cc2d240c727248a175ff67ae7507e35b8d03ee20130a627"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.949282 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" event={"ID":"315c0ef3-a03d-4195-bda6-d08219faeebf","Type":"ContainerStarted","Data":"ade017b966308155848629b0fbe312aca68b9f0d2dff605e2ffff492ca101d65"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.953411 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.953481 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"307aa29c84bcc097a898f8c45bf315abd2c89731f1fe866b8abf1b5fc39f60c3"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.955185 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1fa9ae7a988b8b8635b91e837663fd98f2fafb0ee0a20a76354e204ad1e26991"} Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.973231 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 29 15:03:15 crc kubenswrapper[4767]: I0129 15:03:15.993558 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.019703 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.034114 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.054518 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.073943 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.093278 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.112879 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.132943 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.154027 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.173094 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.192972 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.212964 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.234660 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.253181 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.273617 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.296112 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.313957 4767 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.333624 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.355441 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.374134 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.393946 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.412982 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.450269 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsjjd\" (UniqueName: \"kubernetes.io/projected/d348769a-75fd-4eab-abe7-2dee6bb14f73-kube-api-access-lsjjd\") pod \"console-operator-58897d9998-9dkpb\" (UID: \"d348769a-75fd-4eab-abe7-2dee6bb14f73\") " pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.468827 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8kn8\" (UniqueName: \"kubernetes.io/projected/e542b6c4-ef3e-4673-84ca-528e20243c1c-kube-api-access-p8kn8\") pod \"openshift-controller-manager-operator-756b6f6bc6-mssld\" (UID: \"e542b6c4-ef3e-4673-84ca-528e20243c1c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.471387 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3e628934-9e7a-4c1e-b9fa-018c806f75da-metrics-tls\") pod \"dns-operator-744455d44c-jm5c7\" (UID: \"3e628934-9e7a-4c1e-b9fa-018c806f75da\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.471436 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-node-bootstrap-token\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.471472 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b8a5533d-d711-4fa8-ab7a-a686f77cb578-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qxdzw\" (UID: \"b8a5533d-d711-4fa8-ab7a-a686f77cb578\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.471554 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3affdc50-c158-41f5-a067-c0e3a61541d5-metrics-tls\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.471667 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-certs\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.471694 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-cert\") pod \"ingress-canary-kk65s\" (UID: \"183bee6d-c673-4a5a-a0da-1d3bc44ead6b\") " pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.471801 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3affdc50-c158-41f5-a067-c0e3a61541d5-config-volume\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.475749 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b8a5533d-d711-4fa8-ab7a-a686f77cb578-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-qxdzw\" (UID: \"b8a5533d-d711-4fa8-ab7a-a686f77cb578\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.475796 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3e628934-9e7a-4c1e-b9fa-018c806f75da-metrics-tls\") pod \"dns-operator-744455d44c-jm5c7\" (UID: \"3e628934-9e7a-4c1e-b9fa-018c806f75da\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.476479 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-cert\") pod \"ingress-canary-kk65s\" (UID: \"183bee6d-c673-4a5a-a0da-1d3bc44ead6b\") " pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.498821 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tthx\" (UniqueName: \"kubernetes.io/projected/d8945cab-7833-45ef-b64d-597b5b0db3eb-kube-api-access-8tthx\") pod \"controller-manager-879f6c89f-pvhqb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.508352 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v2cl\" (UniqueName: \"kubernetes.io/projected/393817f3-da00-4bd4-96ae-626b22da8e87-kube-api-access-7v2cl\") pod \"authentication-operator-69f744f599-54tbh\" (UID: \"393817f3-da00-4bd4-96ae-626b22da8e87\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.524761 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.529457 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh8nj\" (UniqueName: \"kubernetes.io/projected/18cf8157-44ad-4a44-8c51-9e41e1d50b52-kube-api-access-jh8nj\") pod \"openshift-apiserver-operator-796bbdcf4f-7qwxt\" (UID: \"18cf8157-44ad-4a44-8c51-9e41e1d50b52\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.530994 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.554475 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mcrc\" (UniqueName: \"kubernetes.io/projected/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-kube-api-access-2mcrc\") pod \"oauth-openshift-558db77b4-4xbzf\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.559576 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.568240 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.569279 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhjvl\" (UniqueName: \"kubernetes.io/projected/737966ab-a636-4868-a7e9-6a2901383f73-kube-api-access-nhjvl\") pod \"downloads-7954f5f757-vbh9z\" (UID: \"737966ab-a636-4868-a7e9-6a2901383f73\") " pod="openshift-console/downloads-7954f5f757-vbh9z" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.588795 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58llh\" (UniqueName: \"kubernetes.io/projected/b9ecb997-b521-4cda-a148-aa4b0f6caa72-kube-api-access-58llh\") pod \"cluster-samples-operator-665b6dd947-njfmj\" (UID: \"b9ecb997-b521-4cda-a148-aa4b0f6caa72\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.594646 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.613507 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.621873 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.632953 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3affdc50-c158-41f5-a067-c0e3a61541d5-metrics-tls\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.633434 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.649019 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3affdc50-c158-41f5-a067-c0e3a61541d5-config-volume\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.651726 4767 request.go:700] Waited for 1.912425038s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.654108 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.673650 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.689241 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-node-bootstrap-token\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.704178 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.718812 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/69d2b93c-da14-4c39-80a2-095dae5a1a42-certs\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.743081 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-54tbh"] Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.763754 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpww7\" (UniqueName: \"kubernetes.io/projected/2356f459-9a7b-4d07-ba43-135788ee3c91-kube-api-access-dpww7\") pod \"apiserver-76f77b778f-qksvg\" (UID: \"2356f459-9a7b-4d07-ba43-135788ee3c91\") " pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.767137 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vbh9z" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.779266 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.793832 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcw8v\" (UniqueName: \"kubernetes.io/projected/1e502829-2ef5-4515-b599-e455100d1e38-kube-api-access-rcw8v\") pod \"migrator-59844c95c7-h26ds\" (UID: \"1e502829-2ef5-4515-b599-e455100d1e38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.794298 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-9dkpb"] Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.796113 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94r67\" (UniqueName: \"kubernetes.io/projected/53138b80-9329-4c54-b607-3df3a694688f-kube-api-access-94r67\") pod \"machine-config-operator-74547568cd-vhcvl\" (UID: \"53138b80-9329-4c54-b607-3df3a694688f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.799946 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.823581 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtmbl\" (UniqueName: \"kubernetes.io/projected/91a22e1c-1a0b-46ac-a8ad-54918a45671d-kube-api-access-qtmbl\") pod \"router-default-5444994796-5g8c7\" (UID: \"91a22e1c-1a0b-46ac-a8ad-54918a45671d\") " pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.853134 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2tfb\" (UniqueName: \"kubernetes.io/projected/ef7f2ec5-eeea-4cf9-87e1-0129b777b90c-kube-api-access-f2tfb\") pod \"etcd-operator-b45778765-fkbz7\" (UID: \"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.853598 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.886101 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55l4m\" (UniqueName: \"kubernetes.io/projected/61181831-03f4-4031-8003-0767b3a18856-kube-api-access-55l4m\") pod \"apiserver-7bbb656c7d-r7dds\" (UID: \"61181831-03f4-4031-8003-0767b3a18856\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.886742 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v8gg\" (UniqueName: \"kubernetes.io/projected/97277fd3-32c2-421d-8a22-99d25e91e14a-kube-api-access-5v8gg\") pod \"olm-operator-6b444d44fb-ztrh7\" (UID: \"97277fd3-32c2-421d-8a22-99d25e91e14a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.903988 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4xbzf"] Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.910349 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7885\" (UniqueName: \"kubernetes.io/projected/1263384a-6a0a-4529-ad99-3162b86e0c20-kube-api-access-m7885\") pod \"kube-storage-version-migrator-operator-b67b599dd-zv56f\" (UID: \"1263384a-6a0a-4529-ad99-3162b86e0c20\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.915091 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.921390 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw794\" (UniqueName: \"kubernetes.io/projected/bd5365ec-5c61-4e64-b3de-d6fd2498084e-kube-api-access-fw794\") pod \"service-ca-9c57cc56f-4rl7s\" (UID: \"bd5365ec-5c61-4e64-b3de-d6fd2498084e\") " pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.922153 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld"] Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.941681 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrp9m\" (UniqueName: \"kubernetes.io/projected/718554cf-a4a8-4a16-b070-64fb54973298-kube-api-access-rrp9m\") pod \"packageserver-d55dfcdfc-8j56k\" (UID: \"718554cf-a4a8-4a16-b070-64fb54973298\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.955648 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc724\" (UniqueName: \"kubernetes.io/projected/4199f360-d546-466f-9dc2-dbd7c0756d93-kube-api-access-qc724\") pod \"route-controller-manager-6576b87f9c-dc9qc\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.959287 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.964192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" event={"ID":"d348769a-75fd-4eab-abe7-2dee6bb14f73","Type":"ContainerStarted","Data":"45c2f9e54475c079e5d5cb23f1dcede8e02c26e76be37d80121ce0e29a108c14"} Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.966365 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" event={"ID":"393817f3-da00-4bd4-96ae-626b22da8e87","Type":"ContainerStarted","Data":"125ead680feee3727f5ad7b05c11bc964c4e1c3530b9375465f7cf31bf6f7a8f"} Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.974946 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj"] Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.976558 4767 csr.go:261] certificate signing request csr-cwln6 is approved, waiting to be issued Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.984221 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.990810 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.992049 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x5ft\" (UniqueName: \"kubernetes.io/projected/e9e8e987-a20f-4f6a-8102-7782d84f88e7-kube-api-access-5x5ft\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.992912 4767 csr.go:257] certificate signing request csr-cwln6 is issued Jan 29 15:03:16 crc kubenswrapper[4767]: I0129 15:03:16.997216 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6vf8\" (UniqueName: \"kubernetes.io/projected/f7932acc-0464-45d2-886a-aefd898e00d0-kube-api-access-r6vf8\") pod \"multus-admission-controller-857f4d67dd-lxfft\" (UID: \"f7932acc-0464-45d2-886a-aefd898e00d0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.014038 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.016177 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.027346 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.037576 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58bph\" (UniqueName: \"kubernetes.io/projected/8c6ee408-b89d-4edf-b18f-d139046e8ccf-kube-api-access-58bph\") pod \"collect-profiles-29494980-lr648\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.046310 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.046361 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx7z6\" (UniqueName: \"kubernetes.io/projected/0c481090-cd6e-428d-bd33-56a4fb452535-kube-api-access-nx7z6\") pod \"package-server-manager-789f6589d5-5vkq5\" (UID: \"0c481090-cd6e-428d-bd33-56a4fb452535\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:17 crc kubenswrapper[4767]: W0129 15:03:17.046491 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91a22e1c_1a0b_46ac_a8ad_54918a45671d.slice/crio-7a8c2f0df8b056e10592532192750b9cfa8ac5659730e9ef11452a65947a3bdc WatchSource:0}: Error finding container 7a8c2f0df8b056e10592532192750b9cfa8ac5659730e9ef11452a65947a3bdc: Status 404 returned error can't find the container with id 7a8c2f0df8b056e10592532192750b9cfa8ac5659730e9ef11452a65947a3bdc Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.055663 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5n27\" (UniqueName: \"kubernetes.io/projected/53b9d35b-24cb-4a15-9f2d-baeac09cdf2a-kube-api-access-x5n27\") pod \"service-ca-operator-777779d784-lpzbl\" (UID: \"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.055820 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.059655 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.059682 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" event={"ID":"e3cbd296-0674-49b6-b9a5-7bb9f6736387","Type":"ContainerStarted","Data":"aebc3878d97cfcc81a965740bcb6ee9d480bae9fd522027a667556c65f1f965c"} Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.059703 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" event={"ID":"315c0ef3-a03d-4195-bda6-d08219faeebf","Type":"ContainerStarted","Data":"051d5af0b029d269bfc86bf6d35f54155611214cf731c1896241e73fbb8b3966"} Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.059716 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" event={"ID":"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3","Type":"ContainerStarted","Data":"53af6631fe1ff4f467a0eb3653399d4e112388f97c6ab6f1f01157e5650781ba"} Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.059728 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" event={"ID":"e542b6c4-ef3e-4673-84ca-528e20243c1c","Type":"ContainerStarted","Data":"db8ee41bd4a67a6a17dfb0d36f9ce21488a4b9c45f181763b0903ee2c7f78e03"} Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.059739 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" event={"ID":"3eca796f-8d95-4e6c-91a0-dd3fb20ea2c2","Type":"ContainerStarted","Data":"887066debed505dd8872391347664c2265dc1c547e0f9c3773bf4c1e3913572a"} Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.072812 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9e8e987-a20f-4f6a-8102-7782d84f88e7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mn2p9\" (UID: \"e9e8e987-a20f-4f6a-8102-7782d84f88e7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.096302 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68fnf\" (UniqueName: \"kubernetes.io/projected/3e628934-9e7a-4c1e-b9fa-018c806f75da-kube-api-access-68fnf\") pod \"dns-operator-744455d44c-jm5c7\" (UID: \"3e628934-9e7a-4c1e-b9fa-018c806f75da\") " pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.121415 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg2sw\" (UniqueName: \"kubernetes.io/projected/183bee6d-c673-4a5a-a0da-1d3bc44ead6b-kube-api-access-pg2sw\") pod \"ingress-canary-kk65s\" (UID: \"183bee6d-c673-4a5a-a0da-1d3bc44ead6b\") " pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.136194 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.140743 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4pvs\" (UniqueName: \"kubernetes.io/projected/b8a5533d-d711-4fa8-ab7a-a686f77cb578-kube-api-access-f4pvs\") pod \"control-plane-machine-set-operator-78cbb6b69f-qxdzw\" (UID: \"b8a5533d-d711-4fa8-ab7a-a686f77cb578\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.141213 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.163755 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r52x\" (UniqueName: \"kubernetes.io/projected/3affdc50-c158-41f5-a067-c0e3a61541d5-kube-api-access-8r52x\") pod \"dns-default-5qnld\" (UID: \"3affdc50-c158-41f5-a067-c0e3a61541d5\") " pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.173278 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhj9r\" (UniqueName: \"kubernetes.io/projected/ed733af7-90b8-481c-86ea-e910a53c36de-kube-api-access-bhj9r\") pod \"csi-hostpathplugin-bk5s7\" (UID: \"ed733af7-90b8-481c-86ea-e910a53c36de\") " pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.198323 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.207220 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.208104 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/05e72201-53f8-42f8-8a74-016cb6506c9f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7hln2\" (UID: \"05e72201-53f8-42f8-8a74-016cb6506c9f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.218023 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8e643b7f-8f80-4aed-a61f-13ffa230b885-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-678bd\" (UID: \"8e643b7f-8f80-4aed-a61f-13ffa230b885\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.219318 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vbh9z"] Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.222556 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt"] Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.234229 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b68b8\" (UniqueName: \"kubernetes.io/projected/e695aba7-f235-4206-9e18-c4cbc877d885-kube-api-access-b68b8\") pod \"catalog-operator-68c6474976-7wdx5\" (UID: \"e695aba7-f235-4206-9e18-c4cbc877d885\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.257883 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hglpd\" (UniqueName: \"kubernetes.io/projected/3d815d7e-754d-454d-8b2b-5891d1fef162-kube-api-access-hglpd\") pod \"machine-config-controller-84d6567774-6px4l\" (UID: \"3d815d7e-754d-454d-8b2b-5891d1fef162\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.267302 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.268411 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.277655 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kk65s" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.278240 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.279760 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp929\" (UniqueName: \"kubernetes.io/projected/69d2b93c-da14-4c39-80a2-095dae5a1a42-kube-api-access-lp929\") pod \"machine-config-server-qq48g\" (UID: \"69d2b93c-da14-4c39-80a2-095dae5a1a42\") " pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:17 crc kubenswrapper[4767]: W0129 15:03:17.285572 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod737966ab_a636_4868_a7e9_6a2901383f73.slice/crio-1527de2712f7b929828500d07d982333b637277fe7812fb5914c035aa125c1de WatchSource:0}: Error finding container 1527de2712f7b929828500d07d982333b637277fe7812fb5914c035aa125c1de: Status 404 returned error can't find the container with id 1527de2712f7b929828500d07d982333b637277fe7812fb5914c035aa125c1de Jan 29 15:03:17 crc kubenswrapper[4767]: W0129 15:03:17.286461 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18cf8157_44ad_4a44_8c51_9e41e1d50b52.slice/crio-0463a774a5fa031af7ec07bcf830ebc172e657e35e80a5c9be56c05f2642670c WatchSource:0}: Error finding container 0463a774a5fa031af7ec07bcf830ebc172e657e35e80a5c9be56c05f2642670c: Status 404 returned error can't find the container with id 0463a774a5fa031af7ec07bcf830ebc172e657e35e80a5c9be56c05f2642670c Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.291674 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.302590 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-qq48g" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320282 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320355 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/565418a4-f20c-4841-92fa-d40804d42980-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320397 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-trusted-ca\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320434 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320517 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6nn8\" (UniqueName: \"kubernetes.io/projected/93676336-30fa-466d-8480-8fecf5d5328e-kube-api-access-b6nn8\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320568 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n6ch\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-kube-api-access-6n6ch\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320603 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320689 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-registry-certificates\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320721 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-bound-sa-token\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320793 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320851 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-registry-tls\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320887 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-config\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320917 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/565418a4-f20c-4841-92fa-d40804d42980-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.320938 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.321408 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:17.821390894 +0000 UTC m=+153.631707933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.325782 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.347449 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.357833 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pvhqb"] Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.373369 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qksvg"] Jan 29 15:03:17 crc kubenswrapper[4767]: W0129 15:03:17.418355 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2356f459_9a7b_4d07_ba43_135788ee3c91.slice/crio-a3da1be4a39e27d5fa903e3a7046811d4f916f3e4096219b9e75bdf29455582c WatchSource:0}: Error finding container a3da1be4a39e27d5fa903e3a7046811d4f916f3e4096219b9e75bdf29455582c: Status 404 returned error can't find the container with id a3da1be4a39e27d5fa903e3a7046811d4f916f3e4096219b9e75bdf29455582c Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.422089 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.422248 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-registry-tls\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.422361 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-config\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.422384 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/565418a4-f20c-4841-92fa-d40804d42980-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.422400 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.422815 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:17.922787132 +0000 UTC m=+153.733104171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.423048 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-config\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.423649 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/565418a4-f20c-4841-92fa-d40804d42980-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.424322 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.424365 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/565418a4-f20c-4841-92fa-d40804d42980-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.424470 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-trusted-ca\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.424521 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.426405 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:17.926386586 +0000 UTC m=+153.736703625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.426696 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.427068 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6nn8\" (UniqueName: \"kubernetes.io/projected/93676336-30fa-466d-8480-8fecf5d5328e-kube-api-access-b6nn8\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.427435 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n6ch\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-kube-api-access-6n6ch\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.428055 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.430590 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-registry-certificates\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.430798 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-bound-sa-token\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.431430 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.436298 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-trusted-ca\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.438867 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-registry-certificates\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.441528 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/565418a4-f20c-4841-92fa-d40804d42980-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.443626 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-registry-tls\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.446142 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.452350 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.459587 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.468087 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.471429 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d344d7a4-05c4-4e38-ac2b-b2eaff9efc14-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9w7fg\" (UID: \"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.476485 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.488678 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.492630 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6nn8\" (UniqueName: \"kubernetes.io/projected/93676336-30fa-466d-8480-8fecf5d5328e-kube-api-access-b6nn8\") pod \"marketplace-operator-79b997595-tsjd8\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.519724 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n6ch\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-kube-api-access-6n6ch\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.532025 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-bound-sa-token\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.534380 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.534735 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.034715924 +0000 UTC m=+153.845032963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.547591 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.638346 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.638681 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.138667625 +0000 UTC m=+153.948984664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.647638 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7"] Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.651310 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds"] Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.689948 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.726563 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-q9hsl" podStartSLOduration=127.726540544 podStartE2EDuration="2m7.726540544s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:17.692165015 +0000 UTC m=+153.502482054" watchObservedRunningTime="2026-01-29 15:03:17.726540544 +0000 UTC m=+153.536857583" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.742436 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.742808 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.242788702 +0000 UTC m=+154.053105741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.847208 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.848598 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.348569946 +0000 UTC m=+154.158886995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.907503 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" podStartSLOduration=127.907480002 podStartE2EDuration="2m7.907480002s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:17.885233251 +0000 UTC m=+153.695550321" watchObservedRunningTime="2026-01-29 15:03:17.907480002 +0000 UTC m=+153.717797041" Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.950155 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.950409 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.450381896 +0000 UTC m=+154.260698935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.950563 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:17 crc kubenswrapper[4767]: E0129 15:03:17.951271 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.451256512 +0000 UTC m=+154.261573551 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.996488 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-29 14:58:16 +0000 UTC, rotation deadline is 2026-11-25 12:01:35.276179165 +0000 UTC Jan 29 15:03:17 crc kubenswrapper[4767]: I0129 15:03:17.996554 4767 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7196h58m17.27962762s for next certificate rotation Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.028544 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.033239 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.052844 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.053342 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.553321849 +0000 UTC m=+154.363638888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.060035 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4rl7s"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.068342 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" event={"ID":"1e502829-2ef5-4515-b599-e455100d1e38","Type":"ContainerStarted","Data":"473e4ea08451fe7c3945b6507598c45292360ae38d099fa7600c56f36a2f8908"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.068443 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fkbz7"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.072958 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" event={"ID":"393817f3-da00-4bd4-96ae-626b22da8e87","Type":"ContainerStarted","Data":"e352e8c7ef65f63c5291add310d9b59300c801b26c075ff991e30a6cd5c67dbb"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.075973 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" event={"ID":"18cf8157-44ad-4a44-8c51-9e41e1d50b52","Type":"ContainerStarted","Data":"9ee0c1263d2d74d6efae95b727bec7cb63386a0a3f56e38d4eb048ce25ad8136"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.076001 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" event={"ID":"18cf8157-44ad-4a44-8c51-9e41e1d50b52","Type":"ContainerStarted","Data":"0463a774a5fa031af7ec07bcf830ebc172e657e35e80a5c9be56c05f2642670c"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.081814 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" event={"ID":"b9ecb997-b521-4cda-a148-aa4b0f6caa72","Type":"ContainerStarted","Data":"1580f98fcdb037517ec663eaef039daa698ade7fd8520bb64589f1d2a0616776"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.082097 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" event={"ID":"b9ecb997-b521-4cda-a148-aa4b0f6caa72","Type":"ContainerStarted","Data":"ec3558e8fb00537f7a6a832a3bbfbfe2ad55246f9af39a0dfc77e5546cc70ec7"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.088791 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" event={"ID":"e542b6c4-ef3e-4673-84ca-528e20243c1c","Type":"ContainerStarted","Data":"32d81aaf71784fc00df269158643df53d1c05fdfa9189999e600e30f1bbe4975"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.089855 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" event={"ID":"97277fd3-32c2-421d-8a22-99d25e91e14a","Type":"ContainerStarted","Data":"43e801c42c911e22f5c992437930170468bb990032a87c7ce2e22a8965e24e21"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.091512 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" event={"ID":"d348769a-75fd-4eab-abe7-2dee6bb14f73","Type":"ContainerStarted","Data":"5d566cf2b37afeb3c1d0322b89eed79c144341c14c634764902ac916d4c75fee"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.091693 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.093166 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" event={"ID":"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3","Type":"ContainerStarted","Data":"6e2a29e8588d3839007b5c69acf683073b4f741648b866403a47a763e3cd9d62"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.093632 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.094562 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" event={"ID":"2356f459-9a7b-4d07-ba43-135788ee3c91","Type":"ContainerStarted","Data":"a3da1be4a39e27d5fa903e3a7046811d4f916f3e4096219b9e75bdf29455582c"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.095959 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" event={"ID":"d8945cab-7833-45ef-b64d-597b5b0db3eb","Type":"ContainerStarted","Data":"98a649888bab06e3e9c736f929bdeae1cdc6bc6a712bc14d54d471a965c203fb"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.097543 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vbh9z" event={"ID":"737966ab-a636-4868-a7e9-6a2901383f73","Type":"ContainerStarted","Data":"1527de2712f7b929828500d07d982333b637277fe7812fb5914c035aa125c1de"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.100238 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qq48g" event={"ID":"69d2b93c-da14-4c39-80a2-095dae5a1a42","Type":"ContainerStarted","Data":"cca34f5211a480cf95805b5347b2faa14254760090b7ef2aa6cdb63a97c1e1ac"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.108686 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5g8c7" event={"ID":"91a22e1c-1a0b-46ac-a8ad-54918a45671d","Type":"ContainerStarted","Data":"6214fdefb616b9384948384beae08daf5025558ccc3ebf3d9d29cd61fb93c138"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.108734 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-5g8c7" event={"ID":"91a22e1c-1a0b-46ac-a8ad-54918a45671d","Type":"ContainerStarted","Data":"7a8c2f0df8b056e10592532192750b9cfa8ac5659730e9ef11452a65947a3bdc"} Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.147732 4767 patch_prober.go:28] interesting pod/console-operator-58897d9998-9dkpb container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.147790 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" podUID="d348769a-75fd-4eab-abe7-2dee6bb14f73" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.149440 4767 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-4xbzf container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.149482 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.156448 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.163853 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.66383684 +0000 UTC m=+154.474153879 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.208413 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-dmgtc" podStartSLOduration=128.208394062 podStartE2EDuration="2m8.208394062s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:18.207796225 +0000 UTC m=+154.018113264" watchObservedRunningTime="2026-01-29 15:03:18.208394062 +0000 UTC m=+154.018711101" Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.259711 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.261722 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.761702426 +0000 UTC m=+154.572019465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.362289 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.363030 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.863018092 +0000 UTC m=+154.673335131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.449106 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wfdcw" podStartSLOduration=128.449089319 podStartE2EDuration="2m8.449089319s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:18.448522463 +0000 UTC m=+154.258839522" watchObservedRunningTime="2026-01-29 15:03:18.449089319 +0000 UTC m=+154.259406358" Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.466230 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.466786 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:18.966762248 +0000 UTC m=+154.777079287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.467589 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.568364 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.569165 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.069130274 +0000 UTC m=+154.879447313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.667563 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.670416 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.670790 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.170768189 +0000 UTC m=+154.981085228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.682058 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.688260 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lxfft"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.701510 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.716840 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kk65s"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.748806 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5qnld"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.750646 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.773333 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.773658 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.27364558 +0000 UTC m=+155.083962619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: W0129 15:03:18.864330 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod183bee6d_c673_4a5a_a0da_1d3bc44ead6b.slice/crio-3684bb2a48b4f04375d43f2096f6170697f50becde2d365dabdc91563dbd0d93 WatchSource:0}: Error finding container 3684bb2a48b4f04375d43f2096f6170697f50becde2d365dabdc91563dbd0d93: Status 404 returned error can't find the container with id 3684bb2a48b4f04375d43f2096f6170697f50becde2d365dabdc91563dbd0d93 Jan 29 15:03:18 crc kubenswrapper[4767]: W0129 15:03:18.865850 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3affdc50_c158_41f5_a067_c0e3a61541d5.slice/crio-a1d8adc638c227e4fdbc8b104224784c1be19192ed2cb50432b4b97e548f6b65 WatchSource:0}: Error finding container a1d8adc638c227e4fdbc8b104224784c1be19192ed2cb50432b4b97e548f6b65: Status 404 returned error can't find the container with id a1d8adc638c227e4fdbc8b104224784c1be19192ed2cb50432b4b97e548f6b65 Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.874141 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.874419 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.37440223 +0000 UTC m=+155.184719269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.893465 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.900244 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.903674 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.930325 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.930439 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.930454 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jm5c7"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.933847 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.937972 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tsjd8"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.957663 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.957819 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.960419 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.973078 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-bk5s7"] Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.975283 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:18 crc kubenswrapper[4767]: E0129 15:03:18.976042 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.476028325 +0000 UTC m=+155.286345364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.976699 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 29 15:03:18 crc kubenswrapper[4767]: I0129 15:03:18.976744 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 29 15:03:19 crc kubenswrapper[4767]: W0129 15:03:19.018065 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9e8e987_a20f_4f6a_8102_7782d84f88e7.slice/crio-df33faf3cf35498c2ce45ef830cfd5d32faef2378527cf909be129c1e0fd094a WatchSource:0}: Error finding container df33faf3cf35498c2ce45ef830cfd5d32faef2378527cf909be129c1e0fd094a: Status 404 returned error can't find the container with id df33faf3cf35498c2ce45ef830cfd5d32faef2378527cf909be129c1e0fd094a Jan 29 15:03:19 crc kubenswrapper[4767]: W0129 15:03:19.045464 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05e72201_53f8_42f8_8a74_016cb6506c9f.slice/crio-3dda5ebc84513825a0fb5025efa8e2f6f7c562ac1f3d34cf996b6d5306dca8d6 WatchSource:0}: Error finding container 3dda5ebc84513825a0fb5025efa8e2f6f7c562ac1f3d34cf996b6d5306dca8d6: Status 404 returned error can't find the container with id 3dda5ebc84513825a0fb5025efa8e2f6f7c562ac1f3d34cf996b6d5306dca8d6 Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.059632 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-89fkh" podStartSLOduration=129.05959999 podStartE2EDuration="2m9.05959999s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.043350772 +0000 UTC m=+154.853667811" watchObservedRunningTime="2026-01-29 15:03:19.05959999 +0000 UTC m=+154.869917029" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.075821 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.076098 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.576067434 +0000 UTC m=+155.386384473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.076484 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.076819 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.576805195 +0000 UTC m=+155.387122234 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.115141 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-qq48g" event={"ID":"69d2b93c-da14-4c39-80a2-095dae5a1a42","Type":"ContainerStarted","Data":"75da68f6b023334b451632cd89a97f0ccb98000042a68c332c2c098edc208c05"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.119125 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" event={"ID":"1e502829-2ef5-4515-b599-e455100d1e38","Type":"ContainerStarted","Data":"fe5f8731632d65fb14d5c2562390f4cecef4684559448463e8c323dd90905c4d"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.121502 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kk65s" event={"ID":"183bee6d-c673-4a5a-a0da-1d3bc44ead6b","Type":"ContainerStarted","Data":"3684bb2a48b4f04375d43f2096f6170697f50becde2d365dabdc91563dbd0d93"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.127824 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" event={"ID":"61181831-03f4-4031-8003-0767b3a18856","Type":"ContainerStarted","Data":"a7941c9ec0d75f56e71b44dda55975f5dea06dda1f2bbacd719bf62233a99268"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.129313 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" event={"ID":"718554cf-a4a8-4a16-b070-64fb54973298","Type":"ContainerStarted","Data":"ed79ac2954505bcb877603899ba38f586585a708c11e380c77ef270de6757349"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.131040 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" event={"ID":"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c","Type":"ContainerStarted","Data":"f5fad3085f7edda8a1084c3a2ecb557258bf07821ad7fbb4fd23877b2af68e34"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.132810 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vbh9z" event={"ID":"737966ab-a636-4868-a7e9-6a2901383f73","Type":"ContainerStarted","Data":"1450cedcdbe9a803ff39414feaa069bc3d9bc3a0382c9fd285a42e9093619633"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.132969 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vbh9z" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.148344 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-vbh9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.148446 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vbh9z" podUID="737966ab-a636-4868-a7e9-6a2901383f73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.150364 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" event={"ID":"53138b80-9329-4c54-b607-3df3a694688f","Type":"ContainerStarted","Data":"df169bd8104f506a6f1b5e6e7d526b96e056f05cf5611329b35a3c84181d8b0f"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.153415 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" event={"ID":"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14","Type":"ContainerStarted","Data":"1c54f748083660fdb961806e40dce8811342a19c49b3312c10671d10754854b0"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.154745 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" event={"ID":"05e72201-53f8-42f8-8a74-016cb6506c9f","Type":"ContainerStarted","Data":"3dda5ebc84513825a0fb5025efa8e2f6f7c562ac1f3d34cf996b6d5306dca8d6"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.161394 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5qnld" event={"ID":"3affdc50-c158-41f5-a067-c0e3a61541d5","Type":"ContainerStarted","Data":"a1d8adc638c227e4fdbc8b104224784c1be19192ed2cb50432b4b97e548f6b65"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.162944 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" event={"ID":"e9e8e987-a20f-4f6a-8102-7782d84f88e7","Type":"ContainerStarted","Data":"df33faf3cf35498c2ce45ef830cfd5d32faef2378527cf909be129c1e0fd094a"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.176024 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" event={"ID":"97277fd3-32c2-421d-8a22-99d25e91e14a","Type":"ContainerStarted","Data":"e1da54425814031ba7e366ecb60f385f0b0d5ef37d582c9b98134c1a442cb707"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.177155 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.178401 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.678384508 +0000 UTC m=+155.488701547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.179199 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" event={"ID":"f7932acc-0464-45d2-886a-aefd898e00d0","Type":"ContainerStarted","Data":"28222208f18c1ce258352018c9cb0558e59fe661f6ec6c4119a82cc727765736"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.187680 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" event={"ID":"0c481090-cd6e-428d-bd33-56a4fb452535","Type":"ContainerStarted","Data":"5f4d1a5b7d6cc60306fea70a62c9ca2df6a55f2c337737e41371cbfba3714318"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.211863 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" event={"ID":"3d815d7e-754d-454d-8b2b-5891d1fef162","Type":"ContainerStarted","Data":"85b6f0d0b4e87423216adaf2b43b6ec1e4b482ed6df88a3aa47987f5085a6028"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.214667 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" event={"ID":"8c6ee408-b89d-4edf-b18f-d139046e8ccf","Type":"ContainerStarted","Data":"2da7c71f614596eeaa00e811b47a02779d502800cd884e961b8415938254fad9"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.220591 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" event={"ID":"1263384a-6a0a-4529-ad99-3162b86e0c20","Type":"ContainerStarted","Data":"851aa9da0b649bb0ea4f81d73d17ee015f591aed6730d8ddaad60e01ef750b30"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.220650 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" event={"ID":"1263384a-6a0a-4529-ad99-3162b86e0c20","Type":"ContainerStarted","Data":"06cd4e0e99724db207efb8b8c25bed56654062ddf68bce99d124a58c4dbf3f6c"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.236211 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" event={"ID":"93676336-30fa-466d-8480-8fecf5d5328e","Type":"ContainerStarted","Data":"bfd32486a6dc7b2d394f284ea303509775e4a792b10d97bdaa901e1fbee645ca"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.278963 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.279962 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.779935651 +0000 UTC m=+155.590252870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.336444 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" podStartSLOduration=129.336425127 podStartE2EDuration="2m9.336425127s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.336356045 +0000 UTC m=+155.146673084" watchObservedRunningTime="2026-01-29 15:03:19.336425127 +0000 UTC m=+155.146742166" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.338955 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" event={"ID":"d8945cab-7833-45ef-b64d-597b5b0db3eb","Type":"ContainerStarted","Data":"3a33af91673ca59609d2347d04b8e8fff8f99c205298f373740637d25e07f242"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.340089 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.342824 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" event={"ID":"b9ecb997-b521-4cda-a148-aa4b0f6caa72","Type":"ContainerStarted","Data":"dfeb93c5d88b410a49f5e6781b4a4f6967d3e893a1d850227c1c8ce2e53c322a"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.347208 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-pvhqb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.347315 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" event={"ID":"b8a5533d-d711-4fa8-ab7a-a686f77cb578","Type":"ContainerStarted","Data":"f51b87402f9666f2de5c7e37019df5097476ad8f20b51394b5959ea6afaa01ea"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.347583 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" podUID="d8945cab-7833-45ef-b64d-597b5b0db3eb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.354637 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" event={"ID":"4199f360-d546-466f-9dc2-dbd7c0756d93","Type":"ContainerStarted","Data":"e38e7c3dc8a0fdd24ccae87419ba1eba9b5b11118e97cc7fcfe32736fb3ff6a8"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.356363 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" event={"ID":"bd5365ec-5c61-4e64-b3de-d6fd2498084e","Type":"ContainerStarted","Data":"1e7f7acdfd2c3d04a6f0872ed7eeaecdc42107ce8cead59936ebc829005301c2"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.360075 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" event={"ID":"8e643b7f-8f80-4aed-a61f-13ffa230b885","Type":"ContainerStarted","Data":"411289adedf70378a3e55936889205e12b247ba9e16b57a55c2cbeda7e46e164"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.364732 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" event={"ID":"3e628934-9e7a-4c1e-b9fa-018c806f75da","Type":"ContainerStarted","Data":"92948ffc8959b5e9bfc3753728ce43d191404f4ac006b158f7976ccad63f3d10"} Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.379649 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.382003 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.881980358 +0000 UTC m=+155.692297397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.414501 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mssld" podStartSLOduration=129.414481373 podStartE2EDuration="2m9.414481373s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.411025124 +0000 UTC m=+155.221342163" watchObservedRunningTime="2026-01-29 15:03:19.414481373 +0000 UTC m=+155.224798412" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.481158 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.481508 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:19.981493632 +0000 UTC m=+155.791810671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.520006 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7qwxt" podStartSLOduration=129.519987429 podStartE2EDuration="2m9.519987429s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.519393782 +0000 UTC m=+155.329710821" watchObservedRunningTime="2026-01-29 15:03:19.519987429 +0000 UTC m=+155.330304468" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.582597 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.582813 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.082781827 +0000 UTC m=+155.893098866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.582866 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" podStartSLOduration=129.582845749 podStartE2EDuration="2m9.582845749s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.575920119 +0000 UTC m=+155.386237158" watchObservedRunningTime="2026-01-29 15:03:19.582845749 +0000 UTC m=+155.393162808" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.583959 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.584629 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.084611099 +0000 UTC m=+155.894928138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.635613 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-9dkpb" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.660067 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.685815 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.686042 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.186017157 +0000 UTC m=+155.996334206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.686108 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.686424 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.186412228 +0000 UTC m=+155.996729267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.737344 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-54tbh" podStartSLOduration=129.737310423 podStartE2EDuration="2m9.737310423s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.7351181 +0000 UTC m=+155.545435139" watchObservedRunningTime="2026-01-29 15:03:19.737310423 +0000 UTC m=+155.547627462" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.787484 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.787970 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.287950611 +0000 UTC m=+156.098267650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.889074 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.890028 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.390000397 +0000 UTC m=+156.200317436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.891975 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vbh9z" podStartSLOduration=129.891946273 podStartE2EDuration="2m9.891946273s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.889596546 +0000 UTC m=+155.699913595" watchObservedRunningTime="2026-01-29 15:03:19.891946273 +0000 UTC m=+155.702263312" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.979043 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:19 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:19 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:19 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.979881 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-5g8c7" podStartSLOduration=129.979845033 podStartE2EDuration="2m9.979845033s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:19.937233217 +0000 UTC m=+155.747550246" watchObservedRunningTime="2026-01-29 15:03:19.979845033 +0000 UTC m=+155.790162072" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.980460 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:19 crc kubenswrapper[4767]: I0129 15:03:19.989976 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:19 crc kubenswrapper[4767]: E0129 15:03:19.990627 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.490609203 +0000 UTC m=+156.300926242 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.095375 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.095925 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.595885513 +0000 UTC m=+156.406202552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.141905 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" podStartSLOduration=130.141868666 podStartE2EDuration="2m10.141868666s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.140567429 +0000 UTC m=+155.950884468" watchObservedRunningTime="2026-01-29 15:03:20.141868666 +0000 UTC m=+155.952185705" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.197394 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.197816 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.697793316 +0000 UTC m=+156.508110355 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.299309 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.299787 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.799770501 +0000 UTC m=+156.610087540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.401108 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.402062 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.902041104 +0000 UTC m=+156.712358153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.402649 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.403167 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:20.903149256 +0000 UTC m=+156.713466295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.407270 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" event={"ID":"718554cf-a4a8-4a16-b070-64fb54973298","Type":"ContainerStarted","Data":"d68aff53c1e64ae1cfb1e588becc357b667b4d4e76ccc69042b936a346cc7f49"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.408840 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.426330 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8j56k container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.426381 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" podUID="718554cf-a4a8-4a16-b070-64fb54973298" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.435037 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" event={"ID":"1e502829-2ef5-4515-b599-e455100d1e38","Type":"ContainerStarted","Data":"1160e31769d0cf90aab81b07a48172871dab0738edf305653ddda719f7eccfad"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.437387 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" event={"ID":"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a","Type":"ContainerStarted","Data":"76a6dbfda7b6f9ab4a5b043bffe46c91bc8f0b8e33ca049f0dc7945f3ce0adfe"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.438748 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" podStartSLOduration=130.438671038 podStartE2EDuration="2m10.438671038s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.435502187 +0000 UTC m=+156.245819226" watchObservedRunningTime="2026-01-29 15:03:20.438671038 +0000 UTC m=+156.248988087" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.440257 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" event={"ID":"ed733af7-90b8-481c-86ea-e910a53c36de","Type":"ContainerStarted","Data":"0db8d914c50aca91eb418480264de26d08ae780a72380c098dd5854a317186db"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.454502 4767 generic.go:334] "Generic (PLEG): container finished" podID="61181831-03f4-4031-8003-0767b3a18856" containerID="0890e3a9417499f60f625a4dde21fa749c5f03afaca3c5203aef480d53c01033" exitCode=0 Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.454592 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" event={"ID":"61181831-03f4-4031-8003-0767b3a18856","Type":"ContainerDied","Data":"0890e3a9417499f60f625a4dde21fa749c5f03afaca3c5203aef480d53c01033"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.460806 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h26ds" podStartSLOduration=130.460786285 podStartE2EDuration="2m10.460786285s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.458854879 +0000 UTC m=+156.269171918" watchObservedRunningTime="2026-01-29 15:03:20.460786285 +0000 UTC m=+156.271103324" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.470077 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" event={"ID":"3d815d7e-754d-454d-8b2b-5891d1fef162","Type":"ContainerStarted","Data":"69832928ce1b2a543d6a88bfec7cbdd9705b3941cf396f646e875f6c1bc72c73"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.476138 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" event={"ID":"ef7f2ec5-eeea-4cf9-87e1-0129b777b90c","Type":"ContainerStarted","Data":"fd037358d1a3f948667ae40a81e4bc5ee071272bc8790893151619acb45a8902"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.504674 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" event={"ID":"b8a5533d-d711-4fa8-ab7a-a686f77cb578","Type":"ContainerStarted","Data":"e722ef53833bf95f042214d624ccd7ac0504561ab7061ae4b022152f6fda0ba2"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.506417 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.510489 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.010468944 +0000 UTC m=+156.820785983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.518430 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" podStartSLOduration=130.518395443 podStartE2EDuration="2m10.518395443s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.510881096 +0000 UTC m=+156.321198135" watchObservedRunningTime="2026-01-29 15:03:20.518395443 +0000 UTC m=+156.328712482" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.523732 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kk65s" event={"ID":"183bee6d-c673-4a5a-a0da-1d3bc44ead6b","Type":"ContainerStarted","Data":"8fd2ae3c11b9b49f9a3785b1a69c6bbd0aad20c36e3616d534e6e4030ba9e4d2"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.536830 4767 generic.go:334] "Generic (PLEG): container finished" podID="2356f459-9a7b-4d07-ba43-135788ee3c91" containerID="e3090fce7ce59f72a19c3ce682908676ff1ecce3af8ed59a593a9bf7d01fdf89" exitCode=0 Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.537090 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" event={"ID":"2356f459-9a7b-4d07-ba43-135788ee3c91","Type":"ContainerDied","Data":"e3090fce7ce59f72a19c3ce682908676ff1ecce3af8ed59a593a9bf7d01fdf89"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.538818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" event={"ID":"bd5365ec-5c61-4e64-b3de-d6fd2498084e","Type":"ContainerStarted","Data":"f0b884d9fd0a1c2bd2597401078fc0bccb9ec644acfc654f7290d016f2d45d2d"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.545118 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" event={"ID":"8e643b7f-8f80-4aed-a61f-13ffa230b885","Type":"ContainerStarted","Data":"5c0347ae291bd8b7e2697793d287f7f292cd35c0760b5052f4cb4b1fe59ac9ce"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.545523 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-qxdzw" podStartSLOduration=130.545498573 podStartE2EDuration="2m10.545498573s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.543306069 +0000 UTC m=+156.353623108" watchObservedRunningTime="2026-01-29 15:03:20.545498573 +0000 UTC m=+156.355815612" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.568377 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" event={"ID":"f7932acc-0464-45d2-886a-aefd898e00d0","Type":"ContainerStarted","Data":"f4d86a876ec05667bbc8a542fb53173075acd646414b0422dd547106f890a6f7"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.593276 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-fkbz7" podStartSLOduration=130.593249627 podStartE2EDuration="2m10.593249627s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.587279075 +0000 UTC m=+156.397596124" watchObservedRunningTime="2026-01-29 15:03:20.593249627 +0000 UTC m=+156.403566666" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.594630 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5qnld" event={"ID":"3affdc50-c158-41f5-a067-c0e3a61541d5","Type":"ContainerStarted","Data":"8f2ebab62c56f551cdf04f38f3940ec1712f8fd5236e6667c7965f6313f72f5a"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.596248 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" event={"ID":"e9e8e987-a20f-4f6a-8102-7782d84f88e7","Type":"ContainerStarted","Data":"3c5c645a22a385cabf59ca93d082b61eca3328b6f9fdefae1188fd42e2a7833c"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.619781 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" event={"ID":"8c6ee408-b89d-4edf-b18f-d139046e8ccf","Type":"ContainerStarted","Data":"91fc9c16f67dbc1338dd226b5538237a4ce9492154c22c51dd7a5241b9b254c7"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.625576 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kk65s" podStartSLOduration=6.625559097 podStartE2EDuration="6.625559097s" podCreationTimestamp="2026-01-29 15:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.624865077 +0000 UTC m=+156.435182116" watchObservedRunningTime="2026-01-29 15:03:20.625559097 +0000 UTC m=+156.435876136" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.627749 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.633773 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" event={"ID":"53138b80-9329-4c54-b607-3df3a694688f","Type":"ContainerStarted","Data":"5ebc9025b4a9b6d606d2aaa0816b60f0743391e6eada07f4809680f3b81af17a"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.633823 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" event={"ID":"53138b80-9329-4c54-b607-3df3a694688f","Type":"ContainerStarted","Data":"0b8b2351b6b3a30578657bdc9ad831447dbbfbc112b0d1cd55e105a4b8aa5dba"} Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.634777 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.134758041 +0000 UTC m=+156.945075080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.639980 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" event={"ID":"e695aba7-f235-4206-9e18-c4cbc877d885","Type":"ContainerStarted","Data":"0e716d3aaf5d46550f242a18e264159417442bee49396bc93bd45c5748473d20"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.640049 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" event={"ID":"e695aba7-f235-4206-9e18-c4cbc877d885","Type":"ContainerStarted","Data":"4e1ea73a5c78728c6be6550eeace5e6ba0193d36550e60dead899478302f7c75"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.642990 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.643494 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-678bd" podStartSLOduration=130.643473022 podStartE2EDuration="2m10.643473022s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.642779402 +0000 UTC m=+156.453096461" watchObservedRunningTime="2026-01-29 15:03:20.643473022 +0000 UTC m=+156.453790061" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.654398 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-7wdx5 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.654542 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" podUID="e695aba7-f235-4206-9e18-c4cbc877d885" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.670137 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-4rl7s" podStartSLOduration=130.670116069 podStartE2EDuration="2m10.670116069s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.667318299 +0000 UTC m=+156.477635338" watchObservedRunningTime="2026-01-29 15:03:20.670116069 +0000 UTC m=+156.480433108" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.681309 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" event={"ID":"4199f360-d546-466f-9dc2-dbd7c0756d93","Type":"ContainerStarted","Data":"40a668810d31e0d3510ce3fe763132005e1e45912feb6712832e44d0c4788947"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.682682 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.691652 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dc9qc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.691725 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.696628 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" event={"ID":"0c481090-cd6e-428d-bd33-56a4fb452535","Type":"ContainerStarted","Data":"cc2b1b1e2da4f5a424e5ab89c60ad3cf26e3b92a553a60f2c403de5caeb08a9e"} Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.697542 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-vbh9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.697625 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vbh9z" podUID="737966ab-a636-4868-a7e9-6a2901383f73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.699137 4767 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-pvhqb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.699192 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" podUID="d8945cab-7833-45ef-b64d-597b5b0db3eb" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.700235 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.730190 4767 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ztrh7 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.730219 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.730714 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" podUID="97277fd3-32c2-421d-8a22-99d25e91e14a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.730311 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.230287601 +0000 UTC m=+157.040604650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.734700 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.738714 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.238697583 +0000 UTC m=+157.049014622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.753025 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zv56f" podStartSLOduration=130.752992024 podStartE2EDuration="2m10.752992024s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.75216087 +0000 UTC m=+156.562477909" watchObservedRunningTime="2026-01-29 15:03:20.752992024 +0000 UTC m=+156.563309073" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.782959 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vhcvl" podStartSLOduration=130.782937156 podStartE2EDuration="2m10.782937156s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.775754149 +0000 UTC m=+156.586071188" watchObservedRunningTime="2026-01-29 15:03:20.782937156 +0000 UTC m=+156.593254195" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.825166 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" podStartSLOduration=130.825142651 podStartE2EDuration="2m10.825142651s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.82304254 +0000 UTC m=+156.633359569" watchObservedRunningTime="2026-01-29 15:03:20.825142651 +0000 UTC m=+156.635459690" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.835864 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.837191 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.337165877 +0000 UTC m=+157.147482916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.848655 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" podStartSLOduration=130.848638807 podStartE2EDuration="2m10.848638807s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.848086831 +0000 UTC m=+156.658403880" watchObservedRunningTime="2026-01-29 15:03:20.848638807 +0000 UTC m=+156.658955836" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.868544 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podStartSLOduration=130.868519199 podStartE2EDuration="2m10.868519199s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.867462609 +0000 UTC m=+156.677779648" watchObservedRunningTime="2026-01-29 15:03:20.868519199 +0000 UTC m=+156.678836248" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.901232 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-qq48g" podStartSLOduration=6.90121207 podStartE2EDuration="6.90121207s" podCreationTimestamp="2026-01-29 15:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.895795304 +0000 UTC m=+156.706112343" watchObservedRunningTime="2026-01-29 15:03:20.90121207 +0000 UTC m=+156.711529119" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.923916 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" podStartSLOduration=130.923887703 podStartE2EDuration="2m10.923887703s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.920981609 +0000 UTC m=+156.731298648" watchObservedRunningTime="2026-01-29 15:03:20.923887703 +0000 UTC m=+156.734204732" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.940740 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:20 crc kubenswrapper[4767]: E0129 15:03:20.941115 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.441104198 +0000 UTC m=+157.251421237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.949970 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-njfmj" podStartSLOduration=130.949953623 podStartE2EDuration="2m10.949953623s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:20.948543102 +0000 UTC m=+156.758860141" watchObservedRunningTime="2026-01-29 15:03:20.949953623 +0000 UTC m=+156.760270662" Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.966029 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:20 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:20 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:20 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:20 crc kubenswrapper[4767]: I0129 15:03:20.966307 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.042340 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.042823 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.542802515 +0000 UTC m=+157.353119554 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.143639 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.144838 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.644822551 +0000 UTC m=+157.455139590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.200715 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ssmkq" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.248211 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.249430 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.749408231 +0000 UTC m=+157.559725270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.349490 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.349827 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.849814131 +0000 UTC m=+157.660131170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.450176 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.450359 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.950333104 +0000 UTC m=+157.760650143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.450990 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.451330 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:21.951321102 +0000 UTC m=+157.761638141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.552458 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.552861 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.052845284 +0000 UTC m=+157.863162323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.654430 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.654875 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.15485248 +0000 UTC m=+157.965169549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.710480 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-lpzbl" event={"ID":"53b9d35b-24cb-4a15-9f2d-baeac09cdf2a","Type":"ContainerStarted","Data":"541917691b301b2a77f995c6f930633b08b2fbc060cbdb5b7044e2f5a470be73"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.713230 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" event={"ID":"e9e8e987-a20f-4f6a-8102-7782d84f88e7","Type":"ContainerStarted","Data":"70e6d5aef18860a96c254a4cd4786d9378c8517bb2bd791ebe963b14d609eec7"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.716432 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" event={"ID":"2356f459-9a7b-4d07-ba43-135788ee3c91","Type":"ContainerStarted","Data":"3709a023036f9f788cbfee2ac9cc4788673b55305a9b1ecfbd41e917208b2b7e"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.716495 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" event={"ID":"2356f459-9a7b-4d07-ba43-135788ee3c91","Type":"ContainerStarted","Data":"98b8d214653f743f6ba9f1e3a79c54b0e9714f93716c956ce04f67923f44c0b1"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.718202 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" event={"ID":"93676336-30fa-466d-8480-8fecf5d5328e","Type":"ContainerStarted","Data":"952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.718825 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.720715 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tsjd8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.720772 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" podUID="93676336-30fa-466d-8480-8fecf5d5328e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.720920 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" event={"ID":"0c481090-cd6e-428d-bd33-56a4fb452535","Type":"ContainerStarted","Data":"29c4a6bfa613eaa9a1b1777e3b635c0c3ac1553250d8a3d50425d547cb328157"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.721041 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.722530 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" event={"ID":"05e72201-53f8-42f8-8a74-016cb6506c9f","Type":"ContainerStarted","Data":"09a9104dbaa5a3f6a267590bd315a435343d99339f67c47ff1d83f022bbdbb2c"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.726072 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" event={"ID":"3e628934-9e7a-4c1e-b9fa-018c806f75da","Type":"ContainerStarted","Data":"e52e13d0e702bf52f4ffdb53be4c579a7ab852875d4f52467078a431d6eb2104"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.726204 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" event={"ID":"3e628934-9e7a-4c1e-b9fa-018c806f75da","Type":"ContainerStarted","Data":"a8e426d86fbb4b958c8c4e648ea422c60460679865f17d48294828d868bdd742"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.727799 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5qnld" event={"ID":"3affdc50-c158-41f5-a067-c0e3a61541d5","Type":"ContainerStarted","Data":"71f0c1f286d5f1f15a85a8c2ac98a946e9ecb01ea6383e282c41b19f06b56de4"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.728097 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.729248 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" event={"ID":"d344d7a4-05c4-4e38-ac2b-b2eaff9efc14","Type":"ContainerStarted","Data":"7b8b3b28a2a63ca62bdb83f9d488c3bf7a518318d97b560446890c7a576ef59a"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.731416 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" event={"ID":"61181831-03f4-4031-8003-0767b3a18856","Type":"ContainerStarted","Data":"c14ae3a4bee9e2a2bb27b6638512d95cb3d8ed0b7da0f55cd743d49821a820e7"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.735013 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn2p9" podStartSLOduration=131.734995936 podStartE2EDuration="2m11.734995936s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.733193054 +0000 UTC m=+157.543510093" watchObservedRunningTime="2026-01-29 15:03:21.734995936 +0000 UTC m=+157.545312975" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.736795 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" event={"ID":"3d815d7e-754d-454d-8b2b-5891d1fef162","Type":"ContainerStarted","Data":"bb55d1e74e7cbbe238d95bf2fb4ce4231e611548c02819d479da25036daa200f"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.740501 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" event={"ID":"f7932acc-0464-45d2-886a-aefd898e00d0","Type":"ContainerStarted","Data":"81f82c8df9e3cbd1c101cc69ed6613f89d7cd288d66c6c9747690b5f6d5774f6"} Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.744262 4767 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-7wdx5 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.744304 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dc9qc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.744364 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.744313 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" podUID="e695aba7-f235-4206-9e18-c4cbc877d885" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.744583 4767 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8j56k container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.744602 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" podUID="718554cf-a4a8-4a16-b070-64fb54973298" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.752454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ztrh7" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.756818 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.758227 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.258208244 +0000 UTC m=+158.068525283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.767431 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7hln2" podStartSLOduration=131.767412439 podStartE2EDuration="2m11.767412439s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.766570745 +0000 UTC m=+157.576887784" watchObservedRunningTime="2026-01-29 15:03:21.767412439 +0000 UTC m=+157.577729478" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.795050 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jm5c7" podStartSLOduration=131.795027704 podStartE2EDuration="2m11.795027704s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.793616563 +0000 UTC m=+157.603933602" watchObservedRunningTime="2026-01-29 15:03:21.795027704 +0000 UTC m=+157.605344743" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.841873 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-5qnld" podStartSLOduration=7.841855151 podStartE2EDuration="7.841855151s" podCreationTimestamp="2026-01-29 15:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.839235466 +0000 UTC m=+157.649552505" watchObservedRunningTime="2026-01-29 15:03:21.841855151 +0000 UTC m=+157.652172190" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.859359 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.866201 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.366182502 +0000 UTC m=+158.176499611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.908996 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" podStartSLOduration=131.908971823 podStartE2EDuration="2m11.908971823s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.906879483 +0000 UTC m=+157.717196522" watchObservedRunningTime="2026-01-29 15:03:21.908971823 +0000 UTC m=+157.719288862" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.910643 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" podStartSLOduration=131.910630601 podStartE2EDuration="2m11.910630601s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.873123541 +0000 UTC m=+157.683440580" watchObservedRunningTime="2026-01-29 15:03:21.910630601 +0000 UTC m=+157.720947640" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.916152 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.916377 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.923299 4767 patch_prober.go:28] interesting pod/apiserver-76f77b778f-qksvg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.17:8443/livez\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.923583 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" podUID="2356f459-9a7b-4d07-ba43-135788ee3c91" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.17:8443/livez\": dial tcp 10.217.0.17:8443: connect: connection refused" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.959096 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9w7fg" podStartSLOduration=131.959074255 podStartE2EDuration="2m11.959074255s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.950747655 +0000 UTC m=+157.761064694" watchObservedRunningTime="2026-01-29 15:03:21.959074255 +0000 UTC m=+157.769391294" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.960842 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.960924 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.460908628 +0000 UTC m=+158.271225667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.961862 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:21 crc kubenswrapper[4767]: E0129 15:03:21.962224 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.462213575 +0000 UTC m=+158.272530664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.969169 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:21 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:21 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:21 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.969236 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:21 crc kubenswrapper[4767]: I0129 15:03:21.988110 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" podStartSLOduration=131.98809452 podStartE2EDuration="2m11.98809452s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:21.98390834 +0000 UTC m=+157.794225379" watchObservedRunningTime="2026-01-29 15:03:21.98809452 +0000 UTC m=+157.798411559" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.027943 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.028461 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.029352 4767 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-r7dds container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.029395 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" podUID="61181831-03f4-4031-8003-0767b3a18856" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.063163 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.063452 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.563436279 +0000 UTC m=+158.373753318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.112079 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" podStartSLOduration=132.112061078 podStartE2EDuration="2m12.112061078s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:22.108290329 +0000 UTC m=+157.918607368" watchObservedRunningTime="2026-01-29 15:03:22.112061078 +0000 UTC m=+157.922378117" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.164339 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.164931 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.664914069 +0000 UTC m=+158.475231108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.265984 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.266260 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.766244795 +0000 UTC m=+158.576561824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.291806 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6px4l" podStartSLOduration=132.29178788 podStartE2EDuration="2m12.29178788s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:22.239488835 +0000 UTC m=+158.049805874" watchObservedRunningTime="2026-01-29 15:03:22.29178788 +0000 UTC m=+158.102104919" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.292989 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-lxfft" podStartSLOduration=132.292981565 podStartE2EDuration="2m12.292981565s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:22.279021483 +0000 UTC m=+158.089338532" watchObservedRunningTime="2026-01-29 15:03:22.292981565 +0000 UTC m=+158.103298614" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.367128 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.367437 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.867425147 +0000 UTC m=+158.677742186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.468269 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.468664 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:22.96864827 +0000 UTC m=+158.778965309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.572844 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.573286 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.073267081 +0000 UTC m=+158.883584120 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.674107 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.674316 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.174290009 +0000 UTC m=+158.984607048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.674689 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.675037 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.17502702 +0000 UTC m=+158.985344129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.747535 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" event={"ID":"ed733af7-90b8-481c-86ea-e910a53c36de","Type":"ContainerStarted","Data":"9288ef05fe9077014fc869335ef9405d4839e47af206efb85bd8b783651552ef"} Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.748288 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tsjd8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.748324 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" podUID="93676336-30fa-466d-8480-8fecf5d5328e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.776836 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.778099 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.278063185 +0000 UTC m=+159.088380224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.818453 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-7wdx5" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.878640 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.880948 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.380936436 +0000 UTC m=+159.191253475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.902698 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.971195 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:22 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:22 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:22 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.971306 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:22 crc kubenswrapper[4767]: I0129 15:03:22.981406 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:22 crc kubenswrapper[4767]: E0129 15:03:22.981780 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.481761598 +0000 UTC m=+159.292078637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.082797 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.084039 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.584026571 +0000 UTC m=+159.394343610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.183794 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.184668 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.684614826 +0000 UTC m=+159.494931865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.285046 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.285547 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.785517119 +0000 UTC m=+159.595834298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.385651 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.385857 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.885829266 +0000 UTC m=+159.696146305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.386018 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.386319 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.88631171 +0000 UTC m=+159.696628749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.487925 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.488110 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.988073819 +0000 UTC m=+159.798390858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.488413 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.488794 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:23.988782719 +0000 UTC m=+159.799099748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.589343 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.589552 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.089507528 +0000 UTC m=+159.899824567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.598134 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8j56k" Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.691002 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.691439 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.191419051 +0000 UTC m=+160.001736090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.753355 4767 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tsjd8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.753394 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" podUID="93676336-30fa-466d-8480-8fecf5d5328e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.792281 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.792614 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.292597413 +0000 UTC m=+160.102914452 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.894778 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.895210 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.395193675 +0000 UTC m=+160.205510714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.970947 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:23 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:23 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:23 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.971063 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:23 crc kubenswrapper[4767]: I0129 15:03:23.996417 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:23 crc kubenswrapper[4767]: E0129 15:03:23.997012 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.496983155 +0000 UTC m=+160.307300194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.098380 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.098763 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.598733023 +0000 UTC m=+160.409050062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.199253 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.199558 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.699541815 +0000 UTC m=+160.509858854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.300681 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.300980 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.800966984 +0000 UTC m=+160.611284023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.385308 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4fjjw"] Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.386666 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.399582 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.403467 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.403614 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.903590177 +0000 UTC m=+160.713907216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.403738 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.404053 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:24.90404472 +0000 UTC m=+160.714361749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.419424 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4fjjw"] Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.504426 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.504662 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.004646506 +0000 UTC m=+160.814963545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.504704 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-catalog-content\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.504733 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztwxr\" (UniqueName: \"kubernetes.io/projected/223f99d9-32ef-4685-9eba-b34f09b337b8-kube-api-access-ztwxr\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.504759 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-utilities\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.504794 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.505354 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.005328505 +0000 UTC m=+160.815645544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.591771 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s7srb"] Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.593252 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.595711 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.605793 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.606056 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-catalog-content\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.606171 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.106065264 +0000 UTC m=+160.916382303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.606292 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztwxr\" (UniqueName: \"kubernetes.io/projected/223f99d9-32ef-4685-9eba-b34f09b337b8-kube-api-access-ztwxr\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.606450 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-utilities\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.606622 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.606626 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-catalog-content\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.606859 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-utilities\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.606968 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.10695991 +0000 UTC m=+160.917276949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.611439 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7srb"] Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.639660 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztwxr\" (UniqueName: \"kubernetes.io/projected/223f99d9-32ef-4685-9eba-b34f09b337b8-kube-api-access-ztwxr\") pod \"certified-operators-4fjjw\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.701217 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.710303 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.710742 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qnth\" (UniqueName: \"kubernetes.io/projected/77c7c371-0113-48f8-b514-0b53a9671aad-kube-api-access-9qnth\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.710854 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-catalog-content\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.710939 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-utilities\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.711106 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.211086967 +0000 UTC m=+161.021404006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.762773 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" event={"ID":"ed733af7-90b8-481c-86ea-e910a53c36de","Type":"ContainerStarted","Data":"26a0665acc45d88264e3253296d664c7d0cccdb3414a6b31f0a0a187ac30333c"} Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.762812 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" event={"ID":"ed733af7-90b8-481c-86ea-e910a53c36de","Type":"ContainerStarted","Data":"4408057de84e8bd029cb84595f76cd4a56712e60506cdb5e8c5020b236c06156"} Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.763732 4767 generic.go:334] "Generic (PLEG): container finished" podID="8c6ee408-b89d-4edf-b18f-d139046e8ccf" containerID="91fc9c16f67dbc1338dd226b5538237a4ce9492154c22c51dd7a5241b9b254c7" exitCode=0 Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.763760 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" event={"ID":"8c6ee408-b89d-4edf-b18f-d139046e8ccf","Type":"ContainerDied","Data":"91fc9c16f67dbc1338dd226b5538237a4ce9492154c22c51dd7a5241b9b254c7"} Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.774112 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x5gw6"] Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.775037 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.800077 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5gw6"] Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.812013 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qnth\" (UniqueName: \"kubernetes.io/projected/77c7c371-0113-48f8-b514-0b53a9671aad-kube-api-access-9qnth\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.812072 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.812100 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-catalog-content\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.812114 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-utilities\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.814988 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.314972626 +0000 UTC m=+161.125289665 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.828706 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-utilities\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.830655 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-catalog-content\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.835749 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qnth\" (UniqueName: \"kubernetes.io/projected/77c7c371-0113-48f8-b514-0b53a9671aad-kube-api-access-9qnth\") pod \"community-operators-s7srb\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.906941 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.913869 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.914112 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-catalog-content\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.914178 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grzjc\" (UniqueName: \"kubernetes.io/projected/46581033-eca1-4394-9760-cf70446cb3ca-kube-api-access-grzjc\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.914208 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-utilities\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:24 crc kubenswrapper[4767]: E0129 15:03:24.914432 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.414405208 +0000 UTC m=+161.224722247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.967868 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:24 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:24 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:24 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.967945 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.983580 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bjs5g"] Jan 29 15:03:24 crc kubenswrapper[4767]: I0129 15:03:24.985146 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.013504 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bjs5g"] Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.017695 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grzjc\" (UniqueName: \"kubernetes.io/projected/46581033-eca1-4394-9760-cf70446cb3ca-kube-api-access-grzjc\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.017766 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-utilities\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.017799 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.017884 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-catalog-content\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.018434 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-catalog-content\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.018993 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-utilities\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.019264 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.519252545 +0000 UTC m=+161.329569604 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.064270 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grzjc\" (UniqueName: \"kubernetes.io/projected/46581033-eca1-4394-9760-cf70446cb3ca-kube-api-access-grzjc\") pod \"certified-operators-x5gw6\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.096320 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.125422 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.125718 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr9ns\" (UniqueName: \"kubernetes.io/projected/e70d6644-4335-4eab-b34c-0530f4a63526-kube-api-access-nr9ns\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.125750 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-catalog-content\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.125809 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-utilities\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.125925 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.625909025 +0000 UTC m=+161.436226064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.186291 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4fjjw"] Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.226794 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.226827 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-utilities\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.226870 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr9ns\" (UniqueName: \"kubernetes.io/projected/e70d6644-4335-4eab-b34c-0530f4a63526-kube-api-access-nr9ns\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.226909 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-catalog-content\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.227853 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-catalog-content\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.228539 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.728520458 +0000 UTC m=+161.538837497 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.228706 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-utilities\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.256229 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr9ns\" (UniqueName: \"kubernetes.io/projected/e70d6644-4335-4eab-b34c-0530f4a63526-kube-api-access-nr9ns\") pod \"community-operators-bjs5g\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.298047 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s7srb"] Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.328337 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.328525 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.828498006 +0000 UTC m=+161.638815045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.328672 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.329068 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.829061062 +0000 UTC m=+161.639378091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: W0129 15:03:25.334763 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77c7c371_0113_48f8_b514_0b53a9671aad.slice/crio-b57c1bb01496f26cd6d6db7020364ea715adea92086e2da0a01519214c9c9401 WatchSource:0}: Error finding container b57c1bb01496f26cd6d6db7020364ea715adea92086e2da0a01519214c9c9401: Status 404 returned error can't find the container with id b57c1bb01496f26cd6d6db7020364ea715adea92086e2da0a01519214c9c9401 Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.339348 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.364337 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.364370 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.374360 4767 patch_prober.go:28] interesting pod/console-f9d7485db-89fkh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.374398 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-89fkh" podUID="861eefc8-26d1-4726-857d-1e44816cdadb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.434717 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.436049 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:25.936033181 +0000 UTC m=+161.746350220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.534009 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5gw6"] Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.540120 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.540847 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:26.040824546 +0000 UTC m=+161.851141585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.641590 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.641918 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:26.141884125 +0000 UTC m=+161.952201164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.656460 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bjs5g"] Jan 29 15:03:25 crc kubenswrapper[4767]: W0129 15:03:25.664041 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode70d6644_4335_4eab_b34c_0530f4a63526.slice/crio-d399bb2031422139cad8fe8e44826c6503ea36f0991bb41ead808908949498d7 WatchSource:0}: Error finding container d399bb2031422139cad8fe8e44826c6503ea36f0991bb41ead808908949498d7: Status 404 returned error can't find the container with id d399bb2031422139cad8fe8e44826c6503ea36f0991bb41ead808908949498d7 Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.743683 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.744116 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:26.244088096 +0000 UTC m=+162.054405135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.775331 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" event={"ID":"ed733af7-90b8-481c-86ea-e910a53c36de","Type":"ContainerStarted","Data":"97be9aa6a2e317410792a6dae62b4a5edd46d3b17a01eda688d53d2a8de7ea50"} Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.777622 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fjjw" event={"ID":"223f99d9-32ef-4685-9eba-b34f09b337b8","Type":"ContainerStarted","Data":"212b10bfda4ef226278c6b0bfa7a744ec4e4fefb8b3fc99173720202c7dfa8d5"} Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.778613 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs5g" event={"ID":"e70d6644-4335-4eab-b34c-0530f4a63526","Type":"ContainerStarted","Data":"d399bb2031422139cad8fe8e44826c6503ea36f0991bb41ead808908949498d7"} Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.779780 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7srb" event={"ID":"77c7c371-0113-48f8-b514-0b53a9671aad","Type":"ContainerStarted","Data":"b57c1bb01496f26cd6d6db7020364ea715adea92086e2da0a01519214c9c9401"} Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.780691 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5gw6" event={"ID":"46581033-eca1-4394-9760-cf70446cb3ca","Type":"ContainerStarted","Data":"aa7955a3fb9297e484db00e71cac6da77f576286e190fb6e9a2fabc256316171"} Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.784469 4767 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.796843 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-bk5s7" podStartSLOduration=11.796824704 podStartE2EDuration="11.796824704s" podCreationTimestamp="2026-01-29 15:03:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:25.795817865 +0000 UTC m=+161.606134904" watchObservedRunningTime="2026-01-29 15:03:25.796824704 +0000 UTC m=+161.607141743" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.845415 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.846265 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:26.346233416 +0000 UTC m=+162.156550455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.953435 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.953798 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:26.453784871 +0000 UTC m=+162.264101910 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.961579 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.962185 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.962252 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.973634 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:25 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:25 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:25 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.973692 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.974132 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 29 15:03:25 crc kubenswrapper[4767]: I0129 15:03:25.977079 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 29 15:03:25 crc kubenswrapper[4767]: E0129 15:03:25.985037 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod223f99d9_32ef_4685_9eba_b34f09b337b8.slice/crio-750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77c7c371_0113_48f8_b514_0b53a9671aad.slice/crio-conmon-ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9.scope\": RecentStats: unable to find data in memory cache]" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.054333 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.054510 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be2fdf7d-ea15-4654-a021-53517b419141-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.054547 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be2fdf7d-ea15-4654-a021-53517b419141-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:26 crc kubenswrapper[4767]: E0129 15:03:26.054687 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-29 15:03:26.554669755 +0000 UTC m=+162.364986794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.122221 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.156253 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be2fdf7d-ea15-4654-a021-53517b419141-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.156318 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be2fdf7d-ea15-4654-a021-53517b419141-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.156372 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.156449 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be2fdf7d-ea15-4654-a021-53517b419141-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:26 crc kubenswrapper[4767]: E0129 15:03:26.156818 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-29 15:03:26.656802894 +0000 UTC m=+162.467120013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rmw5d" (UID: "565418a4-f20c-4841-92fa-d40804d42980") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.174589 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be2fdf7d-ea15-4654-a021-53517b419141-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.198384 4767 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-29T15:03:25.784485469Z","Handler":null,"Name":""} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.201121 4767 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.201155 4767 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.257209 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58bph\" (UniqueName: \"kubernetes.io/projected/8c6ee408-b89d-4edf-b18f-d139046e8ccf-kube-api-access-58bph\") pod \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.257368 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.257402 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c6ee408-b89d-4edf-b18f-d139046e8ccf-secret-volume\") pod \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.257431 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c6ee408-b89d-4edf-b18f-d139046e8ccf-config-volume\") pod \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\" (UID: \"8c6ee408-b89d-4edf-b18f-d139046e8ccf\") " Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.258416 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c6ee408-b89d-4edf-b18f-d139046e8ccf-config-volume" (OuterVolumeSpecName: "config-volume") pod "8c6ee408-b89d-4edf-b18f-d139046e8ccf" (UID: "8c6ee408-b89d-4edf-b18f-d139046e8ccf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.261754 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c6ee408-b89d-4edf-b18f-d139046e8ccf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8c6ee408-b89d-4edf-b18f-d139046e8ccf" (UID: "8c6ee408-b89d-4edf-b18f-d139046e8ccf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.261852 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c6ee408-b89d-4edf-b18f-d139046e8ccf-kube-api-access-58bph" (OuterVolumeSpecName: "kube-api-access-58bph") pod "8c6ee408-b89d-4edf-b18f-d139046e8ccf" (UID: "8c6ee408-b89d-4edf-b18f-d139046e8ccf"). InnerVolumeSpecName "kube-api-access-58bph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.263066 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.359106 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.359499 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58bph\" (UniqueName: \"kubernetes.io/projected/8c6ee408-b89d-4edf-b18f-d139046e8ccf-kube-api-access-58bph\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.359511 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c6ee408-b89d-4edf-b18f-d139046e8ccf-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.359520 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c6ee408-b89d-4edf-b18f-d139046e8ccf-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.362011 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.364834 4767 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.364863 4767 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.381503 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rmw5d\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.516605 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.572336 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 29 15:03:26 crc kubenswrapper[4767]: W0129 15:03:26.577646 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbe2fdf7d_ea15_4654_a021_53517b419141.slice/crio-68b7d4002e0c176d1d124060b4e62f37b2c7e8834e5fa59d0e98dabece253cb2 WatchSource:0}: Error finding container 68b7d4002e0c176d1d124060b4e62f37b2c7e8834e5fa59d0e98dabece253cb2: Status 404 returned error can't find the container with id 68b7d4002e0c176d1d124060b4e62f37b2c7e8834e5fa59d0e98dabece253cb2 Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.578555 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lpc6s"] Jan 29 15:03:26 crc kubenswrapper[4767]: E0129 15:03:26.578747 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c6ee408-b89d-4edf-b18f-d139046e8ccf" containerName="collect-profiles" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.578766 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c6ee408-b89d-4edf-b18f-d139046e8ccf" containerName="collect-profiles" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.578879 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c6ee408-b89d-4edf-b18f-d139046e8ccf" containerName="collect-profiles" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.579572 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.581348 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.582935 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpc6s"] Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.667179 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cpnx\" (UniqueName: \"kubernetes.io/projected/0bed5774-aa4c-4154-9925-6ed6494ca078-kube-api-access-9cpnx\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.667571 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-catalog-content\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.667604 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-utilities\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.732455 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rmw5d"] Jan 29 15:03:26 crc kubenswrapper[4767]: W0129 15:03:26.744581 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod565418a4_f20c_4841_92fa_d40804d42980.slice/crio-40f3383acf2557660bac2f696489e1618b9f5a3df1789c04de3377ce685f485d WatchSource:0}: Error finding container 40f3383acf2557660bac2f696489e1618b9f5a3df1789c04de3377ce685f485d: Status 404 returned error can't find the container with id 40f3383acf2557660bac2f696489e1618b9f5a3df1789c04de3377ce685f485d Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.768334 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-vbh9z container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.768383 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vbh9z" podUID="737966ab-a636-4868-a7e9-6a2901383f73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.768425 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-vbh9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.768481 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vbh9z" podUID="737966ab-a636-4868-a7e9-6a2901383f73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.768686 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cpnx\" (UniqueName: \"kubernetes.io/projected/0bed5774-aa4c-4154-9925-6ed6494ca078-kube-api-access-9cpnx\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.768745 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-catalog-content\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.768776 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-utilities\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.769252 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-catalog-content\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.769572 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-utilities\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.789777 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.791293 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cpnx\" (UniqueName: \"kubernetes.io/projected/0bed5774-aa4c-4154-9925-6ed6494ca078-kube-api-access-9cpnx\") pod \"redhat-marketplace-lpc6s\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.791752 4767 generic.go:334] "Generic (PLEG): container finished" podID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerID="750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3" exitCode=0 Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.791794 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fjjw" event={"ID":"223f99d9-32ef-4685-9eba-b34f09b337b8","Type":"ContainerDied","Data":"750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3"} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.793205 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.793317 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" event={"ID":"565418a4-f20c-4841-92fa-d40804d42980","Type":"ContainerStarted","Data":"40f3383acf2557660bac2f696489e1618b9f5a3df1789c04de3377ce685f485d"} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.800814 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.800816 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648" event={"ID":"8c6ee408-b89d-4edf-b18f-d139046e8ccf","Type":"ContainerDied","Data":"2da7c71f614596eeaa00e811b47a02779d502800cd884e961b8415938254fad9"} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.800864 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2da7c71f614596eeaa00e811b47a02779d502800cd884e961b8415938254fad9" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.803236 4767 generic.go:334] "Generic (PLEG): container finished" podID="e70d6644-4335-4eab-b34c-0530f4a63526" containerID="ce02faf720622b86d315a9e2be68fd2e916092699002784e29846dbc494d08e5" exitCode=0 Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.803311 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs5g" event={"ID":"e70d6644-4335-4eab-b34c-0530f4a63526","Type":"ContainerDied","Data":"ce02faf720622b86d315a9e2be68fd2e916092699002784e29846dbc494d08e5"} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.816551 4767 generic.go:334] "Generic (PLEG): container finished" podID="77c7c371-0113-48f8-b514-0b53a9671aad" containerID="ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9" exitCode=0 Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.816686 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7srb" event={"ID":"77c7c371-0113-48f8-b514-0b53a9671aad","Type":"ContainerDied","Data":"ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9"} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.830230 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"be2fdf7d-ea15-4654-a021-53517b419141","Type":"ContainerStarted","Data":"68b7d4002e0c176d1d124060b4e62f37b2c7e8834e5fa59d0e98dabece253cb2"} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.835203 4767 generic.go:334] "Generic (PLEG): container finished" podID="46581033-eca1-4394-9760-cf70446cb3ca" containerID="1c6456e08602fef5c153f8c5bcdc383169685d98ce5e2ecd260b9d942198f4b8" exitCode=0 Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.837658 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5gw6" event={"ID":"46581033-eca1-4394-9760-cf70446cb3ca","Type":"ContainerDied","Data":"1c6456e08602fef5c153f8c5bcdc383169685d98ce5e2ecd260b9d942198f4b8"} Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.914190 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.931971 4767 patch_prober.go:28] interesting pod/apiserver-76f77b778f-qksvg container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]log ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]etcd ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/generic-apiserver-start-informers ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/max-in-flight-filter ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 29 15:03:26 crc kubenswrapper[4767]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 29 15:03:26 crc kubenswrapper[4767]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/project.openshift.io-projectcache ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-startinformers ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 29 15:03:26 crc kubenswrapper[4767]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 29 15:03:26 crc kubenswrapper[4767]: livez check failed Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.932667 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" podUID="2356f459-9a7b-4d07-ba43-135788ee3c91" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.982392 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.987763 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:26 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:26 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:26 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:26 crc kubenswrapper[4767]: I0129 15:03:26.987997 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.005850 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vffmq"] Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.007110 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.050003 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.050885 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vffmq"] Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.051730 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.059631 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r7dds" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.188610 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bmhw\" (UniqueName: \"kubernetes.io/projected/5370c221-a818-4810-bd42-f824b0f629b1-kube-api-access-8bmhw\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.189056 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-catalog-content\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.189084 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-utilities\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.290790 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bmhw\" (UniqueName: \"kubernetes.io/projected/5370c221-a818-4810-bd42-f824b0f629b1-kube-api-access-8bmhw\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.290846 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-catalog-content\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.290874 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-utilities\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.291539 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-utilities\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.291587 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-catalog-content\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.299783 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpc6s"] Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.321705 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bmhw\" (UniqueName: \"kubernetes.io/projected/5370c221-a818-4810-bd42-f824b0f629b1-kube-api-access-8bmhw\") pod \"redhat-marketplace-vffmq\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.361260 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.578145 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9bblw"] Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.583883 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.588824 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.593152 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9bblw"] Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.654555 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vffmq"] Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.693098 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.699273 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-utilities\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.699375 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t7b7\" (UniqueName: \"kubernetes.io/projected/503997f6-0b05-410b-8011-03f921c6229b-kube-api-access-5t7b7\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.699416 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-catalog-content\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.801583 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-utilities\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.802044 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t7b7\" (UniqueName: \"kubernetes.io/projected/503997f6-0b05-410b-8011-03f921c6229b-kube-api-access-5t7b7\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.802109 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-catalog-content\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.802315 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-utilities\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.802862 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-catalog-content\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.824162 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t7b7\" (UniqueName: \"kubernetes.io/projected/503997f6-0b05-410b-8011-03f921c6229b-kube-api-access-5t7b7\") pod \"redhat-operators-9bblw\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.843829 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" event={"ID":"565418a4-f20c-4841-92fa-d40804d42980","Type":"ContainerStarted","Data":"f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261"} Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.847917 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.856196 4767 generic.go:334] "Generic (PLEG): container finished" podID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerID="d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c" exitCode=0 Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.856461 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpc6s" event={"ID":"0bed5774-aa4c-4154-9925-6ed6494ca078","Type":"ContainerDied","Data":"d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c"} Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.856738 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpc6s" event={"ID":"0bed5774-aa4c-4154-9925-6ed6494ca078","Type":"ContainerStarted","Data":"0de971b7dfb5f72e40142ba5d7131dc8c8dd664786582dbf3e3e5b4a8903c6e9"} Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.859057 4767 generic.go:334] "Generic (PLEG): container finished" podID="be2fdf7d-ea15-4654-a021-53517b419141" containerID="cea2b8262e8b5635c755e7d57c31b1459e1a97a33bb2911ac6f9f30d3565ec5e" exitCode=0 Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.859160 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"be2fdf7d-ea15-4654-a021-53517b419141","Type":"ContainerDied","Data":"cea2b8262e8b5635c755e7d57c31b1459e1a97a33bb2911ac6f9f30d3565ec5e"} Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.876302 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vffmq" event={"ID":"5370c221-a818-4810-bd42-f824b0f629b1","Type":"ContainerStarted","Data":"48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a"} Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.876420 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vffmq" event={"ID":"5370c221-a818-4810-bd42-f824b0f629b1","Type":"ContainerStarted","Data":"fdbed1d3e0eb1650af4803a753844bb44f329d9274101bc4b5672ce6d20b91c8"} Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.912647 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" podStartSLOduration=137.912628456 podStartE2EDuration="2m17.912628456s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:03:27.895277056 +0000 UTC m=+163.705594095" watchObservedRunningTime="2026-01-29 15:03:27.912628456 +0000 UTC m=+163.722945495" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.922958 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.972509 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:27 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:27 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:27 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.972628 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.972726 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xps7s"] Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.973757 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:27 crc kubenswrapper[4767]: I0129 15:03:27.980320 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xps7s"] Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.106010 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-catalog-content\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.106420 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-utilities\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.106602 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29n4h\" (UniqueName: \"kubernetes.io/projected/83954719-42d1-429e-aada-b80c003547bc-kube-api-access-29n4h\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.184641 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9bblw"] Jan 29 15:03:28 crc kubenswrapper[4767]: W0129 15:03:28.193040 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod503997f6_0b05_410b_8011_03f921c6229b.slice/crio-cd4da699fb7d8cbb0096b9259c1ae39308181ffc93b1c16222ba961f447e9a64 WatchSource:0}: Error finding container cd4da699fb7d8cbb0096b9259c1ae39308181ffc93b1c16222ba961f447e9a64: Status 404 returned error can't find the container with id cd4da699fb7d8cbb0096b9259c1ae39308181ffc93b1c16222ba961f447e9a64 Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.207869 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29n4h\" (UniqueName: \"kubernetes.io/projected/83954719-42d1-429e-aada-b80c003547bc-kube-api-access-29n4h\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.207962 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-catalog-content\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.207989 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-utilities\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.208427 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-utilities\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.208461 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-catalog-content\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.233456 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29n4h\" (UniqueName: \"kubernetes.io/projected/83954719-42d1-429e-aada-b80c003547bc-kube-api-access-29n4h\") pod \"redhat-operators-xps7s\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.296945 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.808087 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xps7s"] Jan 29 15:03:28 crc kubenswrapper[4767]: W0129 15:03:28.826632 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83954719_42d1_429e_aada_b80c003547bc.slice/crio-c0a4cfa95cbbee35ba59d470acb475ca92d4fa8d8f5d0e4711932e9f1f9cfe2b WatchSource:0}: Error finding container c0a4cfa95cbbee35ba59d470acb475ca92d4fa8d8f5d0e4711932e9f1f9cfe2b: Status 404 returned error can't find the container with id c0a4cfa95cbbee35ba59d470acb475ca92d4fa8d8f5d0e4711932e9f1f9cfe2b Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.896836 4767 generic.go:334] "Generic (PLEG): container finished" podID="5370c221-a818-4810-bd42-f824b0f629b1" containerID="48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a" exitCode=0 Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.896972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vffmq" event={"ID":"5370c221-a818-4810-bd42-f824b0f629b1","Type":"ContainerDied","Data":"48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a"} Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.900608 4767 generic.go:334] "Generic (PLEG): container finished" podID="503997f6-0b05-410b-8011-03f921c6229b" containerID="c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50" exitCode=0 Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.901123 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9bblw" event={"ID":"503997f6-0b05-410b-8011-03f921c6229b","Type":"ContainerDied","Data":"c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50"} Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.901162 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9bblw" event={"ID":"503997f6-0b05-410b-8011-03f921c6229b","Type":"ContainerStarted","Data":"cd4da699fb7d8cbb0096b9259c1ae39308181ffc93b1c16222ba961f447e9a64"} Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.902951 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xps7s" event={"ID":"83954719-42d1-429e-aada-b80c003547bc","Type":"ContainerStarted","Data":"c0a4cfa95cbbee35ba59d470acb475ca92d4fa8d8f5d0e4711932e9f1f9cfe2b"} Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.966681 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:28 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:28 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:28 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:28 crc kubenswrapper[4767]: I0129 15:03:28.967675 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.058507 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.059233 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.061336 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.063344 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.063634 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.206335 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.225023 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.225117 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.294356 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-5qnld" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.325976 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be2fdf7d-ea15-4654-a021-53517b419141-kubelet-dir\") pod \"be2fdf7d-ea15-4654-a021-53517b419141\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.326168 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be2fdf7d-ea15-4654-a021-53517b419141-kube-api-access\") pod \"be2fdf7d-ea15-4654-a021-53517b419141\" (UID: \"be2fdf7d-ea15-4654-a021-53517b419141\") " Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.326339 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.326432 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.326889 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.326985 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be2fdf7d-ea15-4654-a021-53517b419141-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "be2fdf7d-ea15-4654-a021-53517b419141" (UID: "be2fdf7d-ea15-4654-a021-53517b419141"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.337977 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be2fdf7d-ea15-4654-a021-53517b419141-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "be2fdf7d-ea15-4654-a021-53517b419141" (UID: "be2fdf7d-ea15-4654-a021-53517b419141"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.345149 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.396441 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.428327 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be2fdf7d-ea15-4654-a021-53517b419141-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.428367 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be2fdf7d-ea15-4654-a021-53517b419141-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.609631 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 29 15:03:29 crc kubenswrapper[4767]: W0129 15:03:29.619190 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod22d06f2c_a80b_44d5_a723_f3515e0c5ded.slice/crio-2a50b34ffebab7f536ca249884ae1f2ed799aaba5a13282f5e17a3258bb4c48e WatchSource:0}: Error finding container 2a50b34ffebab7f536ca249884ae1f2ed799aaba5a13282f5e17a3258bb4c48e: Status 404 returned error can't find the container with id 2a50b34ffebab7f536ca249884ae1f2ed799aaba5a13282f5e17a3258bb4c48e Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.924745 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.924741 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"be2fdf7d-ea15-4654-a021-53517b419141","Type":"ContainerDied","Data":"68b7d4002e0c176d1d124060b4e62f37b2c7e8834e5fa59d0e98dabece253cb2"} Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.924853 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68b7d4002e0c176d1d124060b4e62f37b2c7e8834e5fa59d0e98dabece253cb2" Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.927509 4767 generic.go:334] "Generic (PLEG): container finished" podID="83954719-42d1-429e-aada-b80c003547bc" containerID="17b8e4bbcd8cfb2b9dc8cfb69f90347f61b9c74bace2e8f72129e1553e4edf05" exitCode=0 Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.927556 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xps7s" event={"ID":"83954719-42d1-429e-aada-b80c003547bc","Type":"ContainerDied","Data":"17b8e4bbcd8cfb2b9dc8cfb69f90347f61b9c74bace2e8f72129e1553e4edf05"} Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.931364 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"22d06f2c-a80b-44d5-a723-f3515e0c5ded","Type":"ContainerStarted","Data":"2a50b34ffebab7f536ca249884ae1f2ed799aaba5a13282f5e17a3258bb4c48e"} Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.968789 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:29 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:29 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:29 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:29 crc kubenswrapper[4767]: I0129 15:03:29.968842 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:30 crc kubenswrapper[4767]: I0129 15:03:30.963252 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:30 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:30 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:30 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:30 crc kubenswrapper[4767]: I0129 15:03:30.963569 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:30 crc kubenswrapper[4767]: I0129 15:03:30.966435 4767 generic.go:334] "Generic (PLEG): container finished" podID="22d06f2c-a80b-44d5-a723-f3515e0c5ded" containerID="fd7f853ea9a36278f1fa365bacc9b76770fe83dd172818655ae1142c682617ea" exitCode=0 Jan 29 15:03:30 crc kubenswrapper[4767]: I0129 15:03:30.966485 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"22d06f2c-a80b-44d5-a723-f3515e0c5ded","Type":"ContainerDied","Data":"fd7f853ea9a36278f1fa365bacc9b76770fe83dd172818655ae1142c682617ea"} Jan 29 15:03:31 crc kubenswrapper[4767]: I0129 15:03:31.923665 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:31 crc kubenswrapper[4767]: I0129 15:03:31.931091 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-qksvg" Jan 29 15:03:31 crc kubenswrapper[4767]: I0129 15:03:31.962833 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:31 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:31 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:31 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:31 crc kubenswrapper[4767]: I0129 15:03:31.962909 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:32 crc kubenswrapper[4767]: I0129 15:03:32.835482 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:03:32 crc kubenswrapper[4767]: I0129 15:03:32.882913 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/03807b98-087e-4b91-b3ea-f2457587c580-metrics-certs\") pod \"network-metrics-daemon-lhmrn\" (UID: \"03807b98-087e-4b91-b3ea-f2457587c580\") " pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:03:32 crc kubenswrapper[4767]: I0129 15:03:32.963548 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:32 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:32 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:32 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:32 crc kubenswrapper[4767]: I0129 15:03:32.963697 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:33 crc kubenswrapper[4767]: I0129 15:03:33.143866 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-lhmrn" Jan 29 15:03:33 crc kubenswrapper[4767]: I0129 15:03:33.963818 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:33 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:33 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:33 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:33 crc kubenswrapper[4767]: I0129 15:03:33.963870 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:34 crc kubenswrapper[4767]: I0129 15:03:34.961693 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:34 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:34 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:34 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:34 crc kubenswrapper[4767]: I0129 15:03:34.961745 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:35 crc kubenswrapper[4767]: I0129 15:03:35.360030 4767 patch_prober.go:28] interesting pod/console-f9d7485db-89fkh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 29 15:03:35 crc kubenswrapper[4767]: I0129 15:03:35.360087 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-89fkh" podUID="861eefc8-26d1-4726-857d-1e44816cdadb" containerName="console" probeResult="failure" output="Get \"https://10.217.0.6:8443/health\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 29 15:03:35 crc kubenswrapper[4767]: I0129 15:03:35.962393 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:35 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:35 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:35 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:35 crc kubenswrapper[4767]: I0129 15:03:35.962592 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:36 crc kubenswrapper[4767]: I0129 15:03:36.768508 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-vbh9z container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 15:03:36 crc kubenswrapper[4767]: I0129 15:03:36.768556 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vbh9z" podUID="737966ab-a636-4868-a7e9-6a2901383f73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 15:03:36 crc kubenswrapper[4767]: I0129 15:03:36.768508 4767 patch_prober.go:28] interesting pod/downloads-7954f5f757-vbh9z container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Jan 29 15:03:36 crc kubenswrapper[4767]: I0129 15:03:36.768885 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vbh9z" podUID="737966ab-a636-4868-a7e9-6a2901383f73" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Jan 29 15:03:36 crc kubenswrapper[4767]: I0129 15:03:36.962150 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:36 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:36 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:36 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:36 crc kubenswrapper[4767]: I0129 15:03:36.962232 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:37 crc kubenswrapper[4767]: I0129 15:03:37.962609 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:37 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:37 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:37 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:37 crc kubenswrapper[4767]: I0129 15:03:37.962688 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:38 crc kubenswrapper[4767]: I0129 15:03:38.963003 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:38 crc kubenswrapper[4767]: [-]has-synced failed: reason withheld Jan 29 15:03:38 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:38 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:38 crc kubenswrapper[4767]: I0129 15:03:38.963322 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:39 crc kubenswrapper[4767]: I0129 15:03:39.962030 4767 patch_prober.go:28] interesting pod/router-default-5444994796-5g8c7 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 29 15:03:39 crc kubenswrapper[4767]: [+]has-synced ok Jan 29 15:03:39 crc kubenswrapper[4767]: [+]process-running ok Jan 29 15:03:39 crc kubenswrapper[4767]: healthz check failed Jan 29 15:03:39 crc kubenswrapper[4767]: I0129 15:03:39.962122 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-5g8c7" podUID="91a22e1c-1a0b-46ac-a8ad-54918a45671d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 29 15:03:40 crc kubenswrapper[4767]: I0129 15:03:40.049581 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pvhqb"] Jan 29 15:03:40 crc kubenswrapper[4767]: I0129 15:03:40.054284 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" podUID="d8945cab-7833-45ef-b64d-597b5b0db3eb" containerName="controller-manager" containerID="cri-o://3a33af91673ca59609d2347d04b8e8fff8f99c205298f373740637d25e07f242" gracePeriod=30 Jan 29 15:03:40 crc kubenswrapper[4767]: I0129 15:03:40.095151 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc"] Jan 29 15:03:40 crc kubenswrapper[4767]: I0129 15:03:40.095361 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" containerID="cri-o://40a668810d31e0d3510ce3fe763132005e1e45912feb6712832e44d0c4788947" gracePeriod=30 Jan 29 15:03:40 crc kubenswrapper[4767]: I0129 15:03:40.963523 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:40 crc kubenswrapper[4767]: I0129 15:03:40.967128 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-5g8c7" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.079042 4767 generic.go:334] "Generic (PLEG): container finished" podID="d8945cab-7833-45ef-b64d-597b5b0db3eb" containerID="3a33af91673ca59609d2347d04b8e8fff8f99c205298f373740637d25e07f242" exitCode=0 Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.079113 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" event={"ID":"d8945cab-7833-45ef-b64d-597b5b0db3eb","Type":"ContainerDied","Data":"3a33af91673ca59609d2347d04b8e8fff8f99c205298f373740637d25e07f242"} Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.080965 4767 generic.go:334] "Generic (PLEG): container finished" podID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerID="40a668810d31e0d3510ce3fe763132005e1e45912feb6712832e44d0c4788947" exitCode=0 Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.081629 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" event={"ID":"4199f360-d546-466f-9dc2-dbd7c0756d93","Type":"ContainerDied","Data":"40a668810d31e0d3510ce3fe763132005e1e45912feb6712832e44d0c4788947"} Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.862044 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.869541 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.893921 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-bc995945c-42wdw"] Jan 29 15:03:41 crc kubenswrapper[4767]: E0129 15:03:41.894152 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22d06f2c-a80b-44d5-a723-f3515e0c5ded" containerName="pruner" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.894164 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="22d06f2c-a80b-44d5-a723-f3515e0c5ded" containerName="pruner" Jan 29 15:03:41 crc kubenswrapper[4767]: E0129 15:03:41.894178 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8945cab-7833-45ef-b64d-597b5b0db3eb" containerName="controller-manager" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.894184 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8945cab-7833-45ef-b64d-597b5b0db3eb" containerName="controller-manager" Jan 29 15:03:41 crc kubenswrapper[4767]: E0129 15:03:41.894191 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be2fdf7d-ea15-4654-a021-53517b419141" containerName="pruner" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.894199 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="be2fdf7d-ea15-4654-a021-53517b419141" containerName="pruner" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.894301 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8945cab-7833-45ef-b64d-597b5b0db3eb" containerName="controller-manager" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.894320 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="be2fdf7d-ea15-4654-a021-53517b419141" containerName="pruner" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.894330 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="22d06f2c-a80b-44d5-a723-f3515e0c5ded" containerName="pruner" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.894784 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.910403 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-bc995945c-42wdw"] Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.982476 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tthx\" (UniqueName: \"kubernetes.io/projected/d8945cab-7833-45ef-b64d-597b5b0db3eb-kube-api-access-8tthx\") pod \"d8945cab-7833-45ef-b64d-597b5b0db3eb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.982534 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-client-ca\") pod \"d8945cab-7833-45ef-b64d-597b5b0db3eb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.982569 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kubelet-dir\") pod \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.982613 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-proxy-ca-bundles\") pod \"d8945cab-7833-45ef-b64d-597b5b0db3eb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.982635 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-config\") pod \"d8945cab-7833-45ef-b64d-597b5b0db3eb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.982663 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kube-api-access\") pod \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\" (UID: \"22d06f2c-a80b-44d5-a723-f3515e0c5ded\") " Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.982710 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8945cab-7833-45ef-b64d-597b5b0db3eb-serving-cert\") pod \"d8945cab-7833-45ef-b64d-597b5b0db3eb\" (UID: \"d8945cab-7833-45ef-b64d-597b5b0db3eb\") " Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.983247 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "22d06f2c-a80b-44d5-a723-f3515e0c5ded" (UID: "22d06f2c-a80b-44d5-a723-f3515e0c5ded"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.983634 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d8945cab-7833-45ef-b64d-597b5b0db3eb" (UID: "d8945cab-7833-45ef-b64d-597b5b0db3eb"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.983856 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-config" (OuterVolumeSpecName: "config") pod "d8945cab-7833-45ef-b64d-597b5b0db3eb" (UID: "d8945cab-7833-45ef-b64d-597b5b0db3eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.983869 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-client-ca" (OuterVolumeSpecName: "client-ca") pod "d8945cab-7833-45ef-b64d-597b5b0db3eb" (UID: "d8945cab-7833-45ef-b64d-597b5b0db3eb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.988971 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "22d06f2c-a80b-44d5-a723-f3515e0c5ded" (UID: "22d06f2c-a80b-44d5-a723-f3515e0c5ded"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:03:41 crc kubenswrapper[4767]: I0129 15:03:41.989282 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8945cab-7833-45ef-b64d-597b5b0db3eb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d8945cab-7833-45ef-b64d-597b5b0db3eb" (UID: "d8945cab-7833-45ef-b64d-597b5b0db3eb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.006507 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8945cab-7833-45ef-b64d-597b5b0db3eb-kube-api-access-8tthx" (OuterVolumeSpecName: "kube-api-access-8tthx") pod "d8945cab-7833-45ef-b64d-597b5b0db3eb" (UID: "d8945cab-7833-45ef-b64d-597b5b0db3eb"). InnerVolumeSpecName "kube-api-access-8tthx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.094987 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-client-ca\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095047 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfkch\" (UniqueName: \"kubernetes.io/projected/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-kube-api-access-xfkch\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095096 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-config\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095145 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-serving-cert\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095198 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-proxy-ca-bundles\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095285 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8945cab-7833-45ef-b64d-597b5b0db3eb-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095302 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tthx\" (UniqueName: \"kubernetes.io/projected/d8945cab-7833-45ef-b64d-597b5b0db3eb-kube-api-access-8tthx\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095316 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095328 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095678 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095707 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8945cab-7833-45ef-b64d-597b5b0db3eb-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.095717 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22d06f2c-a80b-44d5-a723-f3515e0c5ded-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.101412 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"22d06f2c-a80b-44d5-a723-f3515e0c5ded","Type":"ContainerDied","Data":"2a50b34ffebab7f536ca249884ae1f2ed799aaba5a13282f5e17a3258bb4c48e"} Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.101446 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a50b34ffebab7f536ca249884ae1f2ed799aaba5a13282f5e17a3258bb4c48e" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.101495 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.106346 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" event={"ID":"d8945cab-7833-45ef-b64d-597b5b0db3eb","Type":"ContainerDied","Data":"98a649888bab06e3e9c736f929bdeae1cdc6bc6a712bc14d54d471a965c203fb"} Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.106558 4767 scope.go:117] "RemoveContainer" containerID="3a33af91673ca59609d2347d04b8e8fff8f99c205298f373740637d25e07f242" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.106643 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-pvhqb" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.138615 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pvhqb"] Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.142495 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-pvhqb"] Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.196529 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-proxy-ca-bundles\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.196603 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-client-ca\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.196621 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfkch\" (UniqueName: \"kubernetes.io/projected/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-kube-api-access-xfkch\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.196640 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-config\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.196676 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-serving-cert\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.197856 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-proxy-ca-bundles\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.198503 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-client-ca\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.198777 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-config\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.199907 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-serving-cert\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.219459 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfkch\" (UniqueName: \"kubernetes.io/projected/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-kube-api-access-xfkch\") pod \"controller-manager-bc995945c-42wdw\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.223395 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.806011 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:03:42 crc kubenswrapper[4767]: I0129 15:03:42.806077 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:03:43 crc kubenswrapper[4767]: I0129 15:03:43.024710 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8945cab-7833-45ef-b64d-597b5b0db3eb" path="/var/lib/kubelet/pods/d8945cab-7833-45ef-b64d-597b5b0db3eb/volumes" Jan 29 15:03:45 crc kubenswrapper[4767]: I0129 15:03:45.364767 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:45 crc kubenswrapper[4767]: I0129 15:03:45.369963 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:03:46 crc kubenswrapper[4767]: I0129 15:03:46.523708 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:03:46 crc kubenswrapper[4767]: I0129 15:03:46.785103 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vbh9z" Jan 29 15:03:48 crc kubenswrapper[4767]: I0129 15:03:48.057189 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dc9qc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 15:03:48 crc kubenswrapper[4767]: I0129 15:03:48.058032 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 15:03:53 crc kubenswrapper[4767]: I0129 15:03:53.942515 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 29 15:03:57 crc kubenswrapper[4767]: I0129 15:03:57.352028 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-5vkq5" Jan 29 15:03:58 crc kubenswrapper[4767]: I0129 15:03:58.058016 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dc9qc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: i/o timeout" start-of-body= Jan 29 15:03:58 crc kubenswrapper[4767]: I0129 15:03:58.058118 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: i/o timeout" Jan 29 15:03:59 crc kubenswrapper[4767]: I0129 15:03:59.948625 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-bc995945c-42wdw"] Jan 29 15:04:04 crc kubenswrapper[4767]: E0129 15:04:04.280699 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:04:04 crc kubenswrapper[4767]: E0129 15:04:04.281543 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8bmhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vffmq_openshift-marketplace(5370c221-a818-4810-bd42-f824b0f629b1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:04 crc kubenswrapper[4767]: E0129 15:04:04.283371 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vffmq" podUID="5370c221-a818-4810-bd42-f824b0f629b1" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.695873 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.696644 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.698726 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.699723 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.700935 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.788573 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbe36273-5be1-418f-9e79-b6521de6f60f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.788697 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe36273-5be1-418f-9e79-b6521de6f60f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.889601 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe36273-5be1-418f-9e79-b6521de6f60f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.890018 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbe36273-5be1-418f-9e79-b6521de6f60f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.890125 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbe36273-5be1-418f-9e79-b6521de6f60f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:04 crc kubenswrapper[4767]: I0129 15:04:04.908520 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe36273-5be1-418f-9e79-b6521de6f60f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:05 crc kubenswrapper[4767]: I0129 15:04:05.019360 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:08 crc kubenswrapper[4767]: I0129 15:04:08.057527 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dc9qc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 15:04:08 crc kubenswrapper[4767]: I0129 15:04:08.058110 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.048405 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.049465 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.091315 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.231453 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.231824 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-var-lock\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.232136 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kube-api-access\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.333296 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.333378 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-var-lock\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.333439 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kube-api-access\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.334012 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.334109 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-var-lock\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.358095 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kube-api-access\") pod \"installer-9-crc\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:10 crc kubenswrapper[4767]: I0129 15:04:10.402595 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:04:12 crc kubenswrapper[4767]: I0129 15:04:12.806480 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:04:12 crc kubenswrapper[4767]: I0129 15:04:12.806985 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:04:12 crc kubenswrapper[4767]: I0129 15:04:12.807078 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:04:12 crc kubenswrapper[4767]: I0129 15:04:12.808209 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:04:12 crc kubenswrapper[4767]: I0129 15:04:12.808442 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350" gracePeriod=600 Jan 29 15:04:16 crc kubenswrapper[4767]: E0129 15:04:16.485812 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:04:16 crc kubenswrapper[4767]: E0129 15:04:16.486280 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-29n4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xps7s_openshift-marketplace(83954719-42d1-429e-aada-b80c003547bc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:16 crc kubenswrapper[4767]: E0129 15:04:16.487876 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xps7s" podUID="83954719-42d1-429e-aada-b80c003547bc" Jan 29 15:04:16 crc kubenswrapper[4767]: E0129 15:04:16.490435 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:04:16 crc kubenswrapper[4767]: E0129 15:04:16.490705 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8bmhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vffmq_openshift-marketplace(5370c221-a818-4810-bd42-f824b0f629b1): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:04:16 crc kubenswrapper[4767]: E0129 15:04:16.491923 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-vffmq" podUID="5370c221-a818-4810-bd42-f824b0f629b1" Jan 29 15:04:17 crc kubenswrapper[4767]: I0129 15:04:17.317283 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350" exitCode=0 Jan 29 15:04:17 crc kubenswrapper[4767]: I0129 15:04:17.317370 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350"} Jan 29 15:04:18 crc kubenswrapper[4767]: I0129 15:04:18.056805 4767 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dc9qc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 15:04:18 crc kubenswrapper[4767]: I0129 15:04:18.056957 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 15:04:21 crc kubenswrapper[4767]: E0129 15:04:21.741446 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xps7s" podUID="83954719-42d1-429e-aada-b80c003547bc" Jan 29 15:04:21 crc kubenswrapper[4767]: I0129 15:04:21.803212 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:04:21 crc kubenswrapper[4767]: I0129 15:04:21.853382 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r"] Jan 29 15:04:21 crc kubenswrapper[4767]: E0129 15:04:21.854169 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" Jan 29 15:04:21 crc kubenswrapper[4767]: I0129 15:04:21.854246 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" Jan 29 15:04:21 crc kubenswrapper[4767]: I0129 15:04:21.855412 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" containerName="route-controller-manager" Jan 29 15:04:21 crc kubenswrapper[4767]: I0129 15:04:21.857248 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:21 crc kubenswrapper[4767]: I0129 15:04:21.870226 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r"] Jan 29 15:04:21 crc kubenswrapper[4767]: E0129 15:04:21.943674 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:04:21 crc kubenswrapper[4767]: E0129 15:04:21.943852 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-grzjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-x5gw6_openshift-marketplace(46581033-eca1-4394-9760-cf70446cb3ca): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:21 crc kubenswrapper[4767]: E0129 15:04:21.945052 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-x5gw6" podUID="46581033-eca1-4394-9760-cf70446cb3ca" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001080 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-client-ca\") pod \"4199f360-d546-466f-9dc2-dbd7c0756d93\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001156 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4199f360-d546-466f-9dc2-dbd7c0756d93-serving-cert\") pod \"4199f360-d546-466f-9dc2-dbd7c0756d93\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001250 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-config\") pod \"4199f360-d546-466f-9dc2-dbd7c0756d93\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001291 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc724\" (UniqueName: \"kubernetes.io/projected/4199f360-d546-466f-9dc2-dbd7c0756d93-kube-api-access-qc724\") pod \"4199f360-d546-466f-9dc2-dbd7c0756d93\" (UID: \"4199f360-d546-466f-9dc2-dbd7c0756d93\") " Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001623 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e75e5d01-ff5b-4f84-8267-2a7c5455375c-serving-cert\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001754 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-config\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001822 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-client-ca\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001860 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phshn\" (UniqueName: \"kubernetes.io/projected/e75e5d01-ff5b-4f84-8267-2a7c5455375c-kube-api-access-phshn\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001928 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-client-ca" (OuterVolumeSpecName: "client-ca") pod "4199f360-d546-466f-9dc2-dbd7c0756d93" (UID: "4199f360-d546-466f-9dc2-dbd7c0756d93"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.001999 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-config" (OuterVolumeSpecName: "config") pod "4199f360-d546-466f-9dc2-dbd7c0756d93" (UID: "4199f360-d546-466f-9dc2-dbd7c0756d93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.006549 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4199f360-d546-466f-9dc2-dbd7c0756d93-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4199f360-d546-466f-9dc2-dbd7c0756d93" (UID: "4199f360-d546-466f-9dc2-dbd7c0756d93"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.012169 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4199f360-d546-466f-9dc2-dbd7c0756d93-kube-api-access-qc724" (OuterVolumeSpecName: "kube-api-access-qc724") pod "4199f360-d546-466f-9dc2-dbd7c0756d93" (UID: "4199f360-d546-466f-9dc2-dbd7c0756d93"). InnerVolumeSpecName "kube-api-access-qc724". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.103852 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e75e5d01-ff5b-4f84-8267-2a7c5455375c-serving-cert\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.103980 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-config\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.104036 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-client-ca\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.104081 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phshn\" (UniqueName: \"kubernetes.io/projected/e75e5d01-ff5b-4f84-8267-2a7c5455375c-kube-api-access-phshn\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.104169 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.104191 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4199f360-d546-466f-9dc2-dbd7c0756d93-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.104210 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4199f360-d546-466f-9dc2-dbd7c0756d93-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.104228 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc724\" (UniqueName: \"kubernetes.io/projected/4199f360-d546-466f-9dc2-dbd7c0756d93-kube-api-access-qc724\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.107304 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e75e5d01-ff5b-4f84-8267-2a7c5455375c-serving-cert\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.128080 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-config\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.129013 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-client-ca\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.130979 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phshn\" (UniqueName: \"kubernetes.io/projected/e75e5d01-ff5b-4f84-8267-2a7c5455375c-kube-api-access-phshn\") pod \"route-controller-manager-699d4d85c4-5dm7r\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.174665 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.345422 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" event={"ID":"4199f360-d546-466f-9dc2-dbd7c0756d93","Type":"ContainerDied","Data":"e38e7c3dc8a0fdd24ccae87419ba1eba9b5b11118e97cc7fcfe32736fb3ff6a8"} Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.345451 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc" Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.385459 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc"] Jan 29 15:04:22 crc kubenswrapper[4767]: I0129 15:04:22.388420 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dc9qc"] Jan 29 15:04:23 crc kubenswrapper[4767]: I0129 15:04:23.025658 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4199f360-d546-466f-9dc2-dbd7c0756d93" path="/var/lib/kubelet/pods/4199f360-d546-466f-9dc2-dbd7c0756d93/volumes" Jan 29 15:04:23 crc kubenswrapper[4767]: E0129 15:04:23.194984 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:04:23 crc kubenswrapper[4767]: E0129 15:04:23.195141 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9cpnx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lpc6s_openshift-marketplace(0bed5774-aa4c-4154-9925-6ed6494ca078): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:23 crc kubenswrapper[4767]: E0129 15:04:23.196329 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lpc6s" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" Jan 29 15:04:26 crc kubenswrapper[4767]: E0129 15:04:26.272177 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-x5gw6" podUID="46581033-eca1-4394-9760-cf70446cb3ca" Jan 29 15:04:26 crc kubenswrapper[4767]: E0129 15:04:26.272552 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lpc6s" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" Jan 29 15:04:26 crc kubenswrapper[4767]: I0129 15:04:26.310998 4767 scope.go:117] "RemoveContainer" containerID="40a668810d31e0d3510ce3fe763132005e1e45912feb6712832e44d0c4788947" Jan 29 15:04:26 crc kubenswrapper[4767]: I0129 15:04:26.547734 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-bc995945c-42wdw"] Jan 29 15:04:26 crc kubenswrapper[4767]: W0129 15:04:26.556665 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ed45c53_7b16_4b6e_9b8a_dd8e1e970dc8.slice/crio-1980bf29a61527baa3adeb1500fe8a5486976d529a6bc0b17f2ff51c2ac3c68a WatchSource:0}: Error finding container 1980bf29a61527baa3adeb1500fe8a5486976d529a6bc0b17f2ff51c2ac3c68a: Status 404 returned error can't find the container with id 1980bf29a61527baa3adeb1500fe8a5486976d529a6bc0b17f2ff51c2ac3c68a Jan 29 15:04:26 crc kubenswrapper[4767]: I0129 15:04:26.683380 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-lhmrn"] Jan 29 15:04:26 crc kubenswrapper[4767]: I0129 15:04:26.785463 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r"] Jan 29 15:04:26 crc kubenswrapper[4767]: I0129 15:04:26.788573 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 29 15:04:26 crc kubenswrapper[4767]: I0129 15:04:26.796178 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 29 15:04:26 crc kubenswrapper[4767]: W0129 15:04:26.801710 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podcbe36273_5be1_418f_9e79_b6521de6f60f.slice/crio-ed2aead805d6f7511420e263e0cba470b448ff401b3d9d9a2e8cb9292b44824c WatchSource:0}: Error finding container ed2aead805d6f7511420e263e0cba470b448ff401b3d9d9a2e8cb9292b44824c: Status 404 returned error can't find the container with id ed2aead805d6f7511420e263e0cba470b448ff401b3d9d9a2e8cb9292b44824c Jan 29 15:04:27 crc kubenswrapper[4767]: I0129 15:04:27.375446 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" event={"ID":"e75e5d01-ff5b-4f84-8267-2a7c5455375c","Type":"ContainerStarted","Data":"1e51e74b092dfeb4f9893e969a0a003da07c2cac74197f7c63979a1a2f37d9ae"} Jan 29 15:04:27 crc kubenswrapper[4767]: I0129 15:04:27.376728 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f","Type":"ContainerStarted","Data":"514ca9e58ac272faf454ddf2639c43a81ca40491183f38be85639ff50ca631d1"} Jan 29 15:04:27 crc kubenswrapper[4767]: I0129 15:04:27.378123 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" event={"ID":"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8","Type":"ContainerStarted","Data":"a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522"} Jan 29 15:04:27 crc kubenswrapper[4767]: I0129 15:04:27.378152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" event={"ID":"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8","Type":"ContainerStarted","Data":"1980bf29a61527baa3adeb1500fe8a5486976d529a6bc0b17f2ff51c2ac3c68a"} Jan 29 15:04:27 crc kubenswrapper[4767]: I0129 15:04:27.378929 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"cbe36273-5be1-418f-9e79-b6521de6f60f","Type":"ContainerStarted","Data":"ed2aead805d6f7511420e263e0cba470b448ff401b3d9d9a2e8cb9292b44824c"} Jan 29 15:04:27 crc kubenswrapper[4767]: I0129 15:04:27.379668 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" event={"ID":"03807b98-087e-4b91-b3ea-f2457587c580","Type":"ContainerStarted","Data":"f792a310b2f100ea00f283ce15c0edcd7fcb04eeb8ff90a6788a962bdf044901"} Jan 29 15:04:27 crc kubenswrapper[4767]: I0129 15:04:27.381194 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"2b48f9b4d2a600ad7259d8019bac620bc43f521250d0bb6cd00c64ef387aa854"} Jan 29 15:04:27 crc kubenswrapper[4767]: E0129 15:04:27.591720 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:04:27 crc kubenswrapper[4767]: E0129 15:04:27.592165 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nr9ns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-bjs5g_openshift-marketplace(e70d6644-4335-4eab-b34c-0530f4a63526): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:27 crc kubenswrapper[4767]: E0129 15:04:27.593319 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-bjs5g" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.219765 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.220159 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9qnth,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-s7srb_openshift-marketplace(77c7c371-0113-48f8-b514-0b53a9671aad): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.221952 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-s7srb" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.392184 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" event={"ID":"e75e5d01-ff5b-4f84-8267-2a7c5455375c","Type":"ContainerStarted","Data":"a5f76a6ae079aa93ed2c42fceb958b4d1550d7e050ecba6799369e35c4f47e79"} Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.392691 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.394165 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f","Type":"ContainerStarted","Data":"a4e268ffd57fc7adb97acd6731daf2b1e69cf294b70684f4415f5b1b7592f154"} Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.397032 4767 generic.go:334] "Generic (PLEG): container finished" podID="cbe36273-5be1-418f-9e79-b6521de6f60f" containerID="67e7dea8325a9112e83d5d3c33d8af4a9e9e8625eb01935c57be60fba50f3394" exitCode=0 Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.397168 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"cbe36273-5be1-418f-9e79-b6521de6f60f","Type":"ContainerDied","Data":"67e7dea8325a9112e83d5d3c33d8af4a9e9e8625eb01935c57be60fba50f3394"} Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.399707 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" event={"ID":"03807b98-087e-4b91-b3ea-f2457587c580","Type":"ContainerStarted","Data":"9179852ed68ed42068b7cde7c1a430f9a54fc8cddf0a72e3efd3666e334bef37"} Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.399732 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-lhmrn" event={"ID":"03807b98-087e-4b91-b3ea-f2457587c580","Type":"ContainerStarted","Data":"73aec7f908be3df7875b1ca3a11b3692e9fc6da428b26e3dbb74b430e77e7704"} Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.399804 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" podUID="1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" containerName="controller-manager" containerID="cri-o://a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522" gracePeriod=30 Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.401142 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.402327 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.403964 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-s7srb" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.404084 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-bjs5g" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.419976 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" podStartSLOduration=28.419953442 podStartE2EDuration="28.419953442s" podCreationTimestamp="2026-01-29 15:04:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:04:28.416455678 +0000 UTC m=+224.226772737" watchObservedRunningTime="2026-01-29 15:04:28.419953442 +0000 UTC m=+224.230270491" Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.427425 4767 patch_prober.go:28] interesting pod/controller-manager-bc995945c-42wdw container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:46218->10.217.0.54:8443: read: connection reset by peer" start-of-body= Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.427624 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" podUID="1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:46218->10.217.0.54:8443: read: connection reset by peer" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.443152 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.443282 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5t7b7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-9bblw_openshift-marketplace(503997f6-0b05-410b-8011-03f921c6229b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.444621 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-9bblw" podUID="503997f6-0b05-410b-8011-03f921c6229b" Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.468957 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-lhmrn" podStartSLOduration=198.468936574 podStartE2EDuration="3m18.468936574s" podCreationTimestamp="2026-01-29 15:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:04:28.463199973 +0000 UTC m=+224.273517052" watchObservedRunningTime="2026-01-29 15:04:28.468936574 +0000 UTC m=+224.279253613" Jan 29 15:04:28 crc kubenswrapper[4767]: I0129 15:04:28.538076 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=18.538057847 podStartE2EDuration="18.538057847s" podCreationTimestamp="2026-01-29 15:04:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:04:28.515286977 +0000 UTC m=+224.325604016" watchObservedRunningTime="2026-01-29 15:04:28.538057847 +0000 UTC m=+224.348374886" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.880338 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.880502 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ztwxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4fjjw_openshift-marketplace(223f99d9-32ef-4685-9eba-b34f09b337b8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 29 15:04:28 crc kubenswrapper[4767]: E0129 15:04:28.882235 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4fjjw" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.329517 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.351698 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-645f5bf794-l7w27"] Jan 29 15:04:29 crc kubenswrapper[4767]: E0129 15:04:29.351994 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" containerName="controller-manager" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.352008 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" containerName="controller-manager" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.352142 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" containerName="controller-manager" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.352530 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.369800 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-645f5bf794-l7w27"] Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.405119 4767 generic.go:334] "Generic (PLEG): container finished" podID="1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" containerID="a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522" exitCode=0 Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.405256 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" event={"ID":"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8","Type":"ContainerDied","Data":"a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522"} Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.405329 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.406248 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-bc995945c-42wdw" event={"ID":"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8","Type":"ContainerDied","Data":"1980bf29a61527baa3adeb1500fe8a5486976d529a6bc0b17f2ff51c2ac3c68a"} Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.406275 4767 scope.go:117] "RemoveContainer" containerID="a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522" Jan 29 15:04:29 crc kubenswrapper[4767]: E0129 15:04:29.409867 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4fjjw" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" Jan 29 15:04:29 crc kubenswrapper[4767]: E0129 15:04:29.410080 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-9bblw" podUID="503997f6-0b05-410b-8011-03f921c6229b" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435051 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-serving-cert\") pod \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435144 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-config\") pod \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435213 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfkch\" (UniqueName: \"kubernetes.io/projected/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-kube-api-access-xfkch\") pod \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435239 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-proxy-ca-bundles\") pod \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435303 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-client-ca\") pod \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\" (UID: \"1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8\") " Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435466 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e320cb2b-4888-4eba-a125-b9603ed071c1-serving-cert\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435535 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhvld\" (UniqueName: \"kubernetes.io/projected/e320cb2b-4888-4eba-a125-b9603ed071c1-kube-api-access-lhvld\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435569 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-client-ca\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435594 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-proxy-ca-bundles\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.435623 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-config\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.436377 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" (UID: "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.436454 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-client-ca" (OuterVolumeSpecName: "client-ca") pod "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" (UID: "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.436621 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-config" (OuterVolumeSpecName: "config") pod "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" (UID: "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.436638 4767 scope.go:117] "RemoveContainer" containerID="a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522" Jan 29 15:04:29 crc kubenswrapper[4767]: E0129 15:04:29.437411 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522\": container with ID starting with a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522 not found: ID does not exist" containerID="a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.437448 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522"} err="failed to get container status \"a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522\": rpc error: code = NotFound desc = could not find container \"a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522\": container with ID starting with a237cf456b91065e24547f684df6326cf9df192cece869a0944119f14a8ea522 not found: ID does not exist" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.445223 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" (UID: "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.445259 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-kube-api-access-xfkch" (OuterVolumeSpecName: "kube-api-access-xfkch") pod "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" (UID: "1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8"). InnerVolumeSpecName "kube-api-access-xfkch". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537065 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhvld\" (UniqueName: \"kubernetes.io/projected/e320cb2b-4888-4eba-a125-b9603ed071c1-kube-api-access-lhvld\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537111 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-client-ca\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537134 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-proxy-ca-bundles\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537166 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-config\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537289 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e320cb2b-4888-4eba-a125-b9603ed071c1-serving-cert\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537365 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537378 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537391 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537403 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfkch\" (UniqueName: \"kubernetes.io/projected/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-kube-api-access-xfkch\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.537415 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.538111 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-client-ca\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.538486 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-proxy-ca-bundles\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.538785 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-config\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.544630 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e320cb2b-4888-4eba-a125-b9603ed071c1-serving-cert\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.558811 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhvld\" (UniqueName: \"kubernetes.io/projected/e320cb2b-4888-4eba-a125-b9603ed071c1-kube-api-access-lhvld\") pod \"controller-manager-645f5bf794-l7w27\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.593512 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.638927 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe36273-5be1-418f-9e79-b6521de6f60f-kube-api-access\") pod \"cbe36273-5be1-418f-9e79-b6521de6f60f\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.639440 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbe36273-5be1-418f-9e79-b6521de6f60f-kubelet-dir\") pod \"cbe36273-5be1-418f-9e79-b6521de6f60f\" (UID: \"cbe36273-5be1-418f-9e79-b6521de6f60f\") " Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.639817 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbe36273-5be1-418f-9e79-b6521de6f60f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "cbe36273-5be1-418f-9e79-b6521de6f60f" (UID: "cbe36273-5be1-418f-9e79-b6521de6f60f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.642890 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbe36273-5be1-418f-9e79-b6521de6f60f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "cbe36273-5be1-418f-9e79-b6521de6f60f" (UID: "cbe36273-5be1-418f-9e79-b6521de6f60f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.676460 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.733915 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-bc995945c-42wdw"] Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.736616 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-bc995945c-42wdw"] Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.740823 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cbe36273-5be1-418f-9e79-b6521de6f60f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.740864 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/cbe36273-5be1-418f-9e79-b6521de6f60f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:04:29 crc kubenswrapper[4767]: I0129 15:04:29.902162 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-645f5bf794-l7w27"] Jan 29 15:04:30 crc kubenswrapper[4767]: E0129 15:04:30.020449 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vffmq" podUID="5370c221-a818-4810-bd42-f824b0f629b1" Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.412600 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" event={"ID":"e320cb2b-4888-4eba-a125-b9603ed071c1","Type":"ContainerStarted","Data":"60ffabbd5dc33c41883e31d22cc19b968171c2ee091c7b6a1d1cf4f6fb7a2bf3"} Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.413008 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" event={"ID":"e320cb2b-4888-4eba-a125-b9603ed071c1","Type":"ContainerStarted","Data":"9f227f237510873c38f38278d8705d53bbf0cbe0c503ac6ac0c4cceab36e4ac0"} Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.413052 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.415045 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.415045 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"cbe36273-5be1-418f-9e79-b6521de6f60f","Type":"ContainerDied","Data":"ed2aead805d6f7511420e263e0cba470b448ff401b3d9d9a2e8cb9292b44824c"} Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.415098 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed2aead805d6f7511420e263e0cba470b448ff401b3d9d9a2e8cb9292b44824c" Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.417436 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:04:30 crc kubenswrapper[4767]: I0129 15:04:30.431063 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" podStartSLOduration=31.431050131 podStartE2EDuration="31.431050131s" podCreationTimestamp="2026-01-29 15:03:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:04:30.429769393 +0000 UTC m=+226.240086432" watchObservedRunningTime="2026-01-29 15:04:30.431050131 +0000 UTC m=+226.241367170" Jan 29 15:04:31 crc kubenswrapper[4767]: I0129 15:04:31.025064 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8" path="/var/lib/kubelet/pods/1ed45c53-7b16-4b6e-9b8a-dd8e1e970dc8/volumes" Jan 29 15:04:35 crc kubenswrapper[4767]: E0129 15:04:35.154473 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:04:35 crc kubenswrapper[4767]: E0129 15:04:35.155179 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-29n4h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xps7s_openshift-marketplace(83954719-42d1-429e-aada-b80c003547bc): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:04:35 crc kubenswrapper[4767]: E0129 15:04:35.156479 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-xps7s" podUID="83954719-42d1-429e-aada-b80c003547bc" Jan 29 15:04:39 crc kubenswrapper[4767]: E0129 15:04:39.149824 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:04:39 crc kubenswrapper[4767]: E0129 15:04:39.150591 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-grzjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-x5gw6_openshift-marketplace(46581033-eca1-4394-9760-cf70446cb3ca): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:04:39 crc kubenswrapper[4767]: E0129 15:04:39.151841 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-x5gw6" podUID="46581033-eca1-4394-9760-cf70446cb3ca" Jan 29 15:04:43 crc kubenswrapper[4767]: I0129 15:04:43.268370 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4xbzf"] Jan 29 15:04:43 crc kubenswrapper[4767]: I0129 15:04:43.488931 4767 generic.go:334] "Generic (PLEG): container finished" podID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerID="7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c" exitCode=0 Jan 29 15:04:43 crc kubenswrapper[4767]: I0129 15:04:43.489003 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpc6s" event={"ID":"0bed5774-aa4c-4154-9925-6ed6494ca078","Type":"ContainerDied","Data":"7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c"} Jan 29 15:04:43 crc kubenswrapper[4767]: I0129 15:04:43.493680 4767 generic.go:334] "Generic (PLEG): container finished" podID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerID="143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241" exitCode=0 Jan 29 15:04:43 crc kubenswrapper[4767]: I0129 15:04:43.493707 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fjjw" event={"ID":"223f99d9-32ef-4685-9eba-b34f09b337b8","Type":"ContainerDied","Data":"143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241"} Jan 29 15:04:44 crc kubenswrapper[4767]: I0129 15:04:44.501087 4767 generic.go:334] "Generic (PLEG): container finished" podID="77c7c371-0113-48f8-b514-0b53a9671aad" containerID="7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3" exitCode=0 Jan 29 15:04:44 crc kubenswrapper[4767]: I0129 15:04:44.501107 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7srb" event={"ID":"77c7c371-0113-48f8-b514-0b53a9671aad","Type":"ContainerDied","Data":"7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3"} Jan 29 15:04:44 crc kubenswrapper[4767]: I0129 15:04:44.504703 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9bblw" event={"ID":"503997f6-0b05-410b-8011-03f921c6229b","Type":"ContainerStarted","Data":"dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe"} Jan 29 15:04:44 crc kubenswrapper[4767]: I0129 15:04:44.506865 4767 generic.go:334] "Generic (PLEG): container finished" podID="e70d6644-4335-4eab-b34c-0530f4a63526" containerID="a9adee3e047056ab0ce770a9e5ded56e9ae0af8c5502bd8118c67ce44252986b" exitCode=0 Jan 29 15:04:44 crc kubenswrapper[4767]: I0129 15:04:44.506915 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs5g" event={"ID":"e70d6644-4335-4eab-b34c-0530f4a63526","Type":"ContainerDied","Data":"a9adee3e047056ab0ce770a9e5ded56e9ae0af8c5502bd8118c67ce44252986b"} Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.514238 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs5g" event={"ID":"e70d6644-4335-4eab-b34c-0530f4a63526","Type":"ContainerStarted","Data":"2ef26a0f1410bcc3ef38b7aaac5a472ddc78ec9d4d01500db38598be41533dfc"} Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.517490 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7srb" event={"ID":"77c7c371-0113-48f8-b514-0b53a9671aad","Type":"ContainerStarted","Data":"249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc"} Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.519239 4767 generic.go:334] "Generic (PLEG): container finished" podID="503997f6-0b05-410b-8011-03f921c6229b" containerID="dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe" exitCode=0 Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.519278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9bblw" event={"ID":"503997f6-0b05-410b-8011-03f921c6229b","Type":"ContainerDied","Data":"dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe"} Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.525434 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fjjw" event={"ID":"223f99d9-32ef-4685-9eba-b34f09b337b8","Type":"ContainerStarted","Data":"caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650"} Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.528111 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpc6s" event={"ID":"0bed5774-aa4c-4154-9925-6ed6494ca078","Type":"ContainerStarted","Data":"1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454"} Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.561560 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bjs5g" podStartSLOduration=3.4228986949999998 podStartE2EDuration="1m21.561539798s" podCreationTimestamp="2026-01-29 15:03:24 +0000 UTC" firstStartedPulling="2026-01-29 15:03:26.809423207 +0000 UTC m=+162.619740246" lastFinishedPulling="2026-01-29 15:04:44.9480643 +0000 UTC m=+240.758381349" observedRunningTime="2026-01-29 15:04:45.541762678 +0000 UTC m=+241.352079717" watchObservedRunningTime="2026-01-29 15:04:45.561539798 +0000 UTC m=+241.371856847" Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.579498 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s7srb" podStartSLOduration=3.521347845 podStartE2EDuration="1m21.579477942s" podCreationTimestamp="2026-01-29 15:03:24 +0000 UTC" firstStartedPulling="2026-01-29 15:03:26.824778988 +0000 UTC m=+162.635096027" lastFinishedPulling="2026-01-29 15:04:44.882909085 +0000 UTC m=+240.693226124" observedRunningTime="2026-01-29 15:04:45.578321418 +0000 UTC m=+241.388638457" watchObservedRunningTime="2026-01-29 15:04:45.579477942 +0000 UTC m=+241.389794981" Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.580288 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lpc6s" podStartSLOduration=3.036539914 podStartE2EDuration="1m19.580281826s" podCreationTimestamp="2026-01-29 15:03:26 +0000 UTC" firstStartedPulling="2026-01-29 15:03:27.858126487 +0000 UTC m=+163.668443526" lastFinishedPulling="2026-01-29 15:04:44.401868399 +0000 UTC m=+240.212185438" observedRunningTime="2026-01-29 15:04:45.559831007 +0000 UTC m=+241.370148066" watchObservedRunningTime="2026-01-29 15:04:45.580281826 +0000 UTC m=+241.390598865" Jan 29 15:04:45 crc kubenswrapper[4767]: I0129 15:04:45.594194 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4fjjw" podStartSLOduration=4.046829912 podStartE2EDuration="1m21.594174811s" podCreationTimestamp="2026-01-29 15:03:24 +0000 UTC" firstStartedPulling="2026-01-29 15:03:26.792944352 +0000 UTC m=+162.603261391" lastFinishedPulling="2026-01-29 15:04:44.340289241 +0000 UTC m=+240.150606290" observedRunningTime="2026-01-29 15:04:45.59348858 +0000 UTC m=+241.403805619" watchObservedRunningTime="2026-01-29 15:04:45.594174811 +0000 UTC m=+241.404491850" Jan 29 15:04:46 crc kubenswrapper[4767]: I0129 15:04:46.536451 4767 generic.go:334] "Generic (PLEG): container finished" podID="5370c221-a818-4810-bd42-f824b0f629b1" containerID="3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0" exitCode=0 Jan 29 15:04:46 crc kubenswrapper[4767]: I0129 15:04:46.536528 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vffmq" event={"ID":"5370c221-a818-4810-bd42-f824b0f629b1","Type":"ContainerDied","Data":"3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0"} Jan 29 15:04:46 crc kubenswrapper[4767]: I0129 15:04:46.538949 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9bblw" event={"ID":"503997f6-0b05-410b-8011-03f921c6229b","Type":"ContainerStarted","Data":"285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a"} Jan 29 15:04:46 crc kubenswrapper[4767]: I0129 15:04:46.587873 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9bblw" podStartSLOduration=2.563335425 podStartE2EDuration="1m19.587849566s" podCreationTimestamp="2026-01-29 15:03:27 +0000 UTC" firstStartedPulling="2026-01-29 15:03:28.903304277 +0000 UTC m=+164.713621316" lastFinishedPulling="2026-01-29 15:04:45.927818418 +0000 UTC m=+241.738135457" observedRunningTime="2026-01-29 15:04:46.584739763 +0000 UTC m=+242.395056802" watchObservedRunningTime="2026-01-29 15:04:46.587849566 +0000 UTC m=+242.398166605" Jan 29 15:04:46 crc kubenswrapper[4767]: I0129 15:04:46.915070 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:04:46 crc kubenswrapper[4767]: I0129 15:04:46.915128 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:04:47 crc kubenswrapper[4767]: E0129 15:04:47.030058 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xps7s" podUID="83954719-42d1-429e-aada-b80c003547bc" Jan 29 15:04:47 crc kubenswrapper[4767]: I0129 15:04:47.054171 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:04:47 crc kubenswrapper[4767]: I0129 15:04:47.545618 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vffmq" event={"ID":"5370c221-a818-4810-bd42-f824b0f629b1","Type":"ContainerStarted","Data":"31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc"} Jan 29 15:04:47 crc kubenswrapper[4767]: I0129 15:04:47.566370 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vffmq" podStartSLOduration=2.481135081 podStartE2EDuration="1m21.566342398s" podCreationTimestamp="2026-01-29 15:03:26 +0000 UTC" firstStartedPulling="2026-01-29 15:03:27.895307537 +0000 UTC m=+163.705624576" lastFinishedPulling="2026-01-29 15:04:46.980514854 +0000 UTC m=+242.790831893" observedRunningTime="2026-01-29 15:04:47.564946926 +0000 UTC m=+243.375263975" watchObservedRunningTime="2026-01-29 15:04:47.566342398 +0000 UTC m=+243.376659447" Jan 29 15:04:47 crc kubenswrapper[4767]: I0129 15:04:47.925142 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:04:47 crc kubenswrapper[4767]: I0129 15:04:47.925203 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:04:48 crc kubenswrapper[4767]: I0129 15:04:48.961842 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9bblw" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="registry-server" probeResult="failure" output=< Jan 29 15:04:48 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:04:48 crc kubenswrapper[4767]: > Jan 29 15:04:54 crc kubenswrapper[4767]: I0129 15:04:54.702651 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:04:54 crc kubenswrapper[4767]: I0129 15:04:54.703496 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:04:54 crc kubenswrapper[4767]: I0129 15:04:54.748928 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:04:54 crc kubenswrapper[4767]: I0129 15:04:54.913201 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:04:54 crc kubenswrapper[4767]: I0129 15:04:54.913525 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:04:54 crc kubenswrapper[4767]: I0129 15:04:54.969052 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:04:55 crc kubenswrapper[4767]: E0129 15:04:55.026460 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-x5gw6" podUID="46581033-eca1-4394-9760-cf70446cb3ca" Jan 29 15:04:55 crc kubenswrapper[4767]: I0129 15:04:55.339733 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:04:55 crc kubenswrapper[4767]: I0129 15:04:55.339787 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:04:55 crc kubenswrapper[4767]: I0129 15:04:55.377154 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:04:55 crc kubenswrapper[4767]: I0129 15:04:55.628520 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:04:55 crc kubenswrapper[4767]: I0129 15:04:55.631684 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:04:55 crc kubenswrapper[4767]: I0129 15:04:55.638257 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:04:56 crc kubenswrapper[4767]: I0129 15:04:56.963325 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:04:57 crc kubenswrapper[4767]: I0129 15:04:57.259885 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bjs5g"] Jan 29 15:04:57 crc kubenswrapper[4767]: I0129 15:04:57.362041 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:04:57 crc kubenswrapper[4767]: I0129 15:04:57.362310 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:04:57 crc kubenswrapper[4767]: I0129 15:04:57.406888 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:04:57 crc kubenswrapper[4767]: I0129 15:04:57.597102 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bjs5g" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="registry-server" containerID="cri-o://2ef26a0f1410bcc3ef38b7aaac5a472ddc78ec9d4d01500db38598be41533dfc" gracePeriod=2 Jan 29 15:04:57 crc kubenswrapper[4767]: I0129 15:04:57.658115 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:04:57 crc kubenswrapper[4767]: I0129 15:04:57.964335 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:04:58 crc kubenswrapper[4767]: I0129 15:04:58.015081 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.460426 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vffmq"] Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.619492 4767 generic.go:334] "Generic (PLEG): container finished" podID="e70d6644-4335-4eab-b34c-0530f4a63526" containerID="2ef26a0f1410bcc3ef38b7aaac5a472ddc78ec9d4d01500db38598be41533dfc" exitCode=0 Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.619536 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs5g" event={"ID":"e70d6644-4335-4eab-b34c-0530f4a63526","Type":"ContainerDied","Data":"2ef26a0f1410bcc3ef38b7aaac5a472ddc78ec9d4d01500db38598be41533dfc"} Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.800953 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.901915 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-utilities\") pod \"e70d6644-4335-4eab-b34c-0530f4a63526\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.902200 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-catalog-content\") pod \"e70d6644-4335-4eab-b34c-0530f4a63526\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.902256 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr9ns\" (UniqueName: \"kubernetes.io/projected/e70d6644-4335-4eab-b34c-0530f4a63526-kube-api-access-nr9ns\") pod \"e70d6644-4335-4eab-b34c-0530f4a63526\" (UID: \"e70d6644-4335-4eab-b34c-0530f4a63526\") " Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.902975 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-utilities" (OuterVolumeSpecName: "utilities") pod "e70d6644-4335-4eab-b34c-0530f4a63526" (UID: "e70d6644-4335-4eab-b34c-0530f4a63526"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.910377 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70d6644-4335-4eab-b34c-0530f4a63526-kube-api-access-nr9ns" (OuterVolumeSpecName: "kube-api-access-nr9ns") pod "e70d6644-4335-4eab-b34c-0530f4a63526" (UID: "e70d6644-4335-4eab-b34c-0530f4a63526"). InnerVolumeSpecName "kube-api-access-nr9ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.932795 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-645f5bf794-l7w27"] Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.933275 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" podUID="e320cb2b-4888-4eba-a125-b9603ed071c1" containerName="controller-manager" containerID="cri-o://60ffabbd5dc33c41883e31d22cc19b968171c2ee091c7b6a1d1cf4f6fb7a2bf3" gracePeriod=30 Jan 29 15:04:59 crc kubenswrapper[4767]: I0129 15:04:59.964508 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e70d6644-4335-4eab-b34c-0530f4a63526" (UID: "e70d6644-4335-4eab-b34c-0530f4a63526"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.003994 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.004029 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr9ns\" (UniqueName: \"kubernetes.io/projected/e70d6644-4335-4eab-b34c-0530f4a63526-kube-api-access-nr9ns\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.004043 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70d6644-4335-4eab-b34c-0530f4a63526-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.030634 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r"] Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.030882 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" podUID="e75e5d01-ff5b-4f84-8267-2a7c5455375c" containerName="route-controller-manager" containerID="cri-o://a5f76a6ae079aa93ed2c42fceb958b4d1550d7e050ecba6799369e35c4f47e79" gracePeriod=30 Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.627149 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bjs5g" event={"ID":"e70d6644-4335-4eab-b34c-0530f4a63526","Type":"ContainerDied","Data":"d399bb2031422139cad8fe8e44826c6503ea36f0991bb41ead808908949498d7"} Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.627199 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bjs5g" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.627210 4767 scope.go:117] "RemoveContainer" containerID="2ef26a0f1410bcc3ef38b7aaac5a472ddc78ec9d4d01500db38598be41533dfc" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.633861 4767 generic.go:334] "Generic (PLEG): container finished" podID="e320cb2b-4888-4eba-a125-b9603ed071c1" containerID="60ffabbd5dc33c41883e31d22cc19b968171c2ee091c7b6a1d1cf4f6fb7a2bf3" exitCode=0 Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.633943 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" event={"ID":"e320cb2b-4888-4eba-a125-b9603ed071c1","Type":"ContainerDied","Data":"60ffabbd5dc33c41883e31d22cc19b968171c2ee091c7b6a1d1cf4f6fb7a2bf3"} Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.635968 4767 generic.go:334] "Generic (PLEG): container finished" podID="e75e5d01-ff5b-4f84-8267-2a7c5455375c" containerID="a5f76a6ae079aa93ed2c42fceb958b4d1550d7e050ecba6799369e35c4f47e79" exitCode=0 Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.636284 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vffmq" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="registry-server" containerID="cri-o://31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc" gracePeriod=2 Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.636440 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" event={"ID":"e75e5d01-ff5b-4f84-8267-2a7c5455375c","Type":"ContainerDied","Data":"a5f76a6ae079aa93ed2c42fceb958b4d1550d7e050ecba6799369e35c4f47e79"} Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.648854 4767 scope.go:117] "RemoveContainer" containerID="a9adee3e047056ab0ce770a9e5ded56e9ae0af8c5502bd8118c67ce44252986b" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.665117 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bjs5g"] Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.667953 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bjs5g"] Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.674514 4767 scope.go:117] "RemoveContainer" containerID="ce02faf720622b86d315a9e2be68fd2e916092699002784e29846dbc494d08e5" Jan 29 15:05:00 crc kubenswrapper[4767]: I0129 15:05:00.994636 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.026645 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" path="/var/lib/kubelet/pods/e70d6644-4335-4eab-b34c-0530f4a63526/volumes" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.068126 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.073341 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.123340 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phshn\" (UniqueName: \"kubernetes.io/projected/e75e5d01-ff5b-4f84-8267-2a7c5455375c-kube-api-access-phshn\") pod \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.123483 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e75e5d01-ff5b-4f84-8267-2a7c5455375c-serving-cert\") pod \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.123525 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-config\") pod \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.123546 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-client-ca\") pod \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\" (UID: \"e75e5d01-ff5b-4f84-8267-2a7c5455375c\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.124736 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-client-ca" (OuterVolumeSpecName: "client-ca") pod "e75e5d01-ff5b-4f84-8267-2a7c5455375c" (UID: "e75e5d01-ff5b-4f84-8267-2a7c5455375c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.125592 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-config" (OuterVolumeSpecName: "config") pod "e75e5d01-ff5b-4f84-8267-2a7c5455375c" (UID: "e75e5d01-ff5b-4f84-8267-2a7c5455375c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.129996 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75e5d01-ff5b-4f84-8267-2a7c5455375c-kube-api-access-phshn" (OuterVolumeSpecName: "kube-api-access-phshn") pod "e75e5d01-ff5b-4f84-8267-2a7c5455375c" (UID: "e75e5d01-ff5b-4f84-8267-2a7c5455375c"). InnerVolumeSpecName "kube-api-access-phshn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.130003 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e75e5d01-ff5b-4f84-8267-2a7c5455375c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e75e5d01-ff5b-4f84-8267-2a7c5455375c" (UID: "e75e5d01-ff5b-4f84-8267-2a7c5455375c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225117 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhvld\" (UniqueName: \"kubernetes.io/projected/e320cb2b-4888-4eba-a125-b9603ed071c1-kube-api-access-lhvld\") pod \"e320cb2b-4888-4eba-a125-b9603ed071c1\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225184 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-client-ca\") pod \"e320cb2b-4888-4eba-a125-b9603ed071c1\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225222 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-proxy-ca-bundles\") pod \"e320cb2b-4888-4eba-a125-b9603ed071c1\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225250 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-config\") pod \"e320cb2b-4888-4eba-a125-b9603ed071c1\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225317 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bmhw\" (UniqueName: \"kubernetes.io/projected/5370c221-a818-4810-bd42-f824b0f629b1-kube-api-access-8bmhw\") pod \"5370c221-a818-4810-bd42-f824b0f629b1\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225345 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-catalog-content\") pod \"5370c221-a818-4810-bd42-f824b0f629b1\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225394 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e320cb2b-4888-4eba-a125-b9603ed071c1-serving-cert\") pod \"e320cb2b-4888-4eba-a125-b9603ed071c1\" (UID: \"e320cb2b-4888-4eba-a125-b9603ed071c1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225453 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-utilities\") pod \"5370c221-a818-4810-bd42-f824b0f629b1\" (UID: \"5370c221-a818-4810-bd42-f824b0f629b1\") " Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225704 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e75e5d01-ff5b-4f84-8267-2a7c5455375c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225720 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225731 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e75e5d01-ff5b-4f84-8267-2a7c5455375c-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225742 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phshn\" (UniqueName: \"kubernetes.io/projected/e75e5d01-ff5b-4f84-8267-2a7c5455375c-kube-api-access-phshn\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.225946 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-client-ca" (OuterVolumeSpecName: "client-ca") pod "e320cb2b-4888-4eba-a125-b9603ed071c1" (UID: "e320cb2b-4888-4eba-a125-b9603ed071c1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.226689 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-utilities" (OuterVolumeSpecName: "utilities") pod "5370c221-a818-4810-bd42-f824b0f629b1" (UID: "5370c221-a818-4810-bd42-f824b0f629b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.227090 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e320cb2b-4888-4eba-a125-b9603ed071c1" (UID: "e320cb2b-4888-4eba-a125-b9603ed071c1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.226969 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-config" (OuterVolumeSpecName: "config") pod "e320cb2b-4888-4eba-a125-b9603ed071c1" (UID: "e320cb2b-4888-4eba-a125-b9603ed071c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.228334 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5370c221-a818-4810-bd42-f824b0f629b1-kube-api-access-8bmhw" (OuterVolumeSpecName: "kube-api-access-8bmhw") pod "5370c221-a818-4810-bd42-f824b0f629b1" (UID: "5370c221-a818-4810-bd42-f824b0f629b1"). InnerVolumeSpecName "kube-api-access-8bmhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.229879 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e320cb2b-4888-4eba-a125-b9603ed071c1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e320cb2b-4888-4eba-a125-b9603ed071c1" (UID: "e320cb2b-4888-4eba-a125-b9603ed071c1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.230330 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e320cb2b-4888-4eba-a125-b9603ed071c1-kube-api-access-lhvld" (OuterVolumeSpecName: "kube-api-access-lhvld") pod "e320cb2b-4888-4eba-a125-b9603ed071c1" (UID: "e320cb2b-4888-4eba-a125-b9603ed071c1"). InnerVolumeSpecName "kube-api-access-lhvld". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.258215 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5370c221-a818-4810-bd42-f824b0f629b1" (UID: "5370c221-a818-4810-bd42-f824b0f629b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326611 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhvld\" (UniqueName: \"kubernetes.io/projected/e320cb2b-4888-4eba-a125-b9603ed071c1-kube-api-access-lhvld\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326850 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326867 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326879 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e320cb2b-4888-4eba-a125-b9603ed071c1-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326937 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bmhw\" (UniqueName: \"kubernetes.io/projected/5370c221-a818-4810-bd42-f824b0f629b1-kube-api-access-8bmhw\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326962 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326977 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e320cb2b-4888-4eba-a125-b9603ed071c1-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.326991 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5370c221-a818-4810-bd42-f824b0f629b1-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.427337 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr"] Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.427797 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe36273-5be1-418f-9e79-b6521de6f60f" containerName="pruner" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.427834 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe36273-5be1-418f-9e79-b6521de6f60f" containerName="pruner" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.427863 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="extract-utilities" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.427881 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="extract-utilities" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.427954 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="extract-content" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.427975 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="extract-content" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.428004 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e320cb2b-4888-4eba-a125-b9603ed071c1" containerName="controller-manager" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428021 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e320cb2b-4888-4eba-a125-b9603ed071c1" containerName="controller-manager" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.428048 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="extract-content" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428066 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="extract-content" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.428096 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="registry-server" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428114 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="registry-server" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.428140 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="registry-server" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428155 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="registry-server" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.428177 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="extract-utilities" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428193 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="extract-utilities" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.428221 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75e5d01-ff5b-4f84-8267-2a7c5455375c" containerName="route-controller-manager" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428239 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75e5d01-ff5b-4f84-8267-2a7c5455375c" containerName="route-controller-manager" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428479 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e75e5d01-ff5b-4f84-8267-2a7c5455375c" containerName="route-controller-manager" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428513 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70d6644-4335-4eab-b34c-0530f4a63526" containerName="registry-server" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428540 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e320cb2b-4888-4eba-a125-b9603ed071c1" containerName="controller-manager" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428569 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbe36273-5be1-418f-9e79-b6521de6f60f" containerName="pruner" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.428586 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5370c221-a818-4810-bd42-f824b0f629b1" containerName="registry-server" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.429432 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.453703 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr"] Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.529631 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-client-ca\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.529724 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d2e8f49-4924-4616-8ba7-ed0225358ace-serving-cert\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.529771 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-config\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.529811 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmn9q\" (UniqueName: \"kubernetes.io/projected/1d2e8f49-4924-4616-8ba7-ed0225358ace-kube-api-access-rmn9q\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.631302 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-client-ca\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.631402 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d2e8f49-4924-4616-8ba7-ed0225358ace-serving-cert\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.631456 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-config\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.631499 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmn9q\" (UniqueName: \"kubernetes.io/projected/1d2e8f49-4924-4616-8ba7-ed0225358ace-kube-api-access-rmn9q\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.633838 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-client-ca\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.634404 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-config\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.638420 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d2e8f49-4924-4616-8ba7-ed0225358ace-serving-cert\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.647051 4767 generic.go:334] "Generic (PLEG): container finished" podID="5370c221-a818-4810-bd42-f824b0f629b1" containerID="31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc" exitCode=0 Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.647119 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vffmq" event={"ID":"5370c221-a818-4810-bd42-f824b0f629b1","Type":"ContainerDied","Data":"31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc"} Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.647150 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vffmq" event={"ID":"5370c221-a818-4810-bd42-f824b0f629b1","Type":"ContainerDied","Data":"fdbed1d3e0eb1650af4803a753844bb44f329d9274101bc4b5672ce6d20b91c8"} Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.647167 4767 scope.go:117] "RemoveContainer" containerID="31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.647185 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vffmq" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.651661 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" event={"ID":"e320cb2b-4888-4eba-a125-b9603ed071c1","Type":"ContainerDied","Data":"9f227f237510873c38f38278d8705d53bbf0cbe0c503ac6ac0c4cceab36e4ac0"} Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.651730 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-645f5bf794-l7w27" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.653679 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" event={"ID":"e75e5d01-ff5b-4f84-8267-2a7c5455375c","Type":"ContainerDied","Data":"1e51e74b092dfeb4f9893e969a0a003da07c2cac74197f7c63979a1a2f37d9ae"} Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.653764 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.656819 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmn9q\" (UniqueName: \"kubernetes.io/projected/1d2e8f49-4924-4616-8ba7-ed0225358ace-kube-api-access-rmn9q\") pod \"route-controller-manager-54f4c4644f-w9ckr\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.664628 4767 scope.go:117] "RemoveContainer" containerID="3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.692067 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-645f5bf794-l7w27"] Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.701501 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-645f5bf794-l7w27"] Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.701736 4767 scope.go:117] "RemoveContainer" containerID="48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.708628 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vffmq"] Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.712574 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vffmq"] Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.718109 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r"] Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.721630 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-699d4d85c4-5dm7r"] Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.728186 4767 scope.go:117] "RemoveContainer" containerID="31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.728813 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc\": container with ID starting with 31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc not found: ID does not exist" containerID="31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.728854 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc"} err="failed to get container status \"31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc\": rpc error: code = NotFound desc = could not find container \"31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc\": container with ID starting with 31d434341b5ab6b4842791d1a559f24f1675d69253f702c41e4dc731f7e290bc not found: ID does not exist" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.728881 4767 scope.go:117] "RemoveContainer" containerID="3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.729305 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0\": container with ID starting with 3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0 not found: ID does not exist" containerID="3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.729413 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0"} err="failed to get container status \"3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0\": rpc error: code = NotFound desc = could not find container \"3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0\": container with ID starting with 3a2a5a70cae2db9d85a6a93771ff5b152a8a533eaf017843c4dfc37145034ae0 not found: ID does not exist" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.729507 4767 scope.go:117] "RemoveContainer" containerID="48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a" Jan 29 15:05:01 crc kubenswrapper[4767]: E0129 15:05:01.729947 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a\": container with ID starting with 48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a not found: ID does not exist" containerID="48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.730012 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a"} err="failed to get container status \"48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a\": rpc error: code = NotFound desc = could not find container \"48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a\": container with ID starting with 48731b9d35a40670dd0b43cfc1cd1c8892ebb2665918e3ec263667137d62868a not found: ID does not exist" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.730057 4767 scope.go:117] "RemoveContainer" containerID="60ffabbd5dc33c41883e31d22cc19b968171c2ee091c7b6a1d1cf4f6fb7a2bf3" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.747175 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.756475 4767 scope.go:117] "RemoveContainer" containerID="a5f76a6ae079aa93ed2c42fceb958b4d1550d7e050ecba6799369e35c4f47e79" Jan 29 15:05:01 crc kubenswrapper[4767]: I0129 15:05:01.965284 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr"] Jan 29 15:05:01 crc kubenswrapper[4767]: W0129 15:05:01.968169 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d2e8f49_4924_4616_8ba7_ed0225358ace.slice/crio-71a8f2bd3135fcbba97ad0c3b518b0ca235ce23b8afcb157b8c2467de950e564 WatchSource:0}: Error finding container 71a8f2bd3135fcbba97ad0c3b518b0ca235ce23b8afcb157b8c2467de950e564: Status 404 returned error can't find the container with id 71a8f2bd3135fcbba97ad0c3b518b0ca235ce23b8afcb157b8c2467de950e564 Jan 29 15:05:02 crc kubenswrapper[4767]: I0129 15:05:02.668003 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" event={"ID":"1d2e8f49-4924-4616-8ba7-ed0225358ace","Type":"ContainerStarted","Data":"bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1"} Jan 29 15:05:02 crc kubenswrapper[4767]: I0129 15:05:02.668162 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" event={"ID":"1d2e8f49-4924-4616-8ba7-ed0225358ace","Type":"ContainerStarted","Data":"71a8f2bd3135fcbba97ad0c3b518b0ca235ce23b8afcb157b8c2467de950e564"} Jan 29 15:05:02 crc kubenswrapper[4767]: I0129 15:05:02.669583 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:02 crc kubenswrapper[4767]: I0129 15:05:02.677984 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:05:02 crc kubenswrapper[4767]: I0129 15:05:02.690333 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" podStartSLOduration=2.690309069 podStartE2EDuration="2.690309069s" podCreationTimestamp="2026-01-29 15:05:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:05:02.687348681 +0000 UTC m=+258.497665730" watchObservedRunningTime="2026-01-29 15:05:02.690309069 +0000 UTC m=+258.500626118" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.026661 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5370c221-a818-4810-bd42-f824b0f629b1" path="/var/lib/kubelet/pods/5370c221-a818-4810-bd42-f824b0f629b1/volumes" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.028813 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e320cb2b-4888-4eba-a125-b9603ed071c1" path="/var/lib/kubelet/pods/e320cb2b-4888-4eba-a125-b9603ed071c1/volumes" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.030464 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e75e5d01-ff5b-4f84-8267-2a7c5455375c" path="/var/lib/kubelet/pods/e75e5d01-ff5b-4f84-8267-2a7c5455375c/volumes" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.438799 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g"] Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.440080 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.443604 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.443727 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.445146 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.446046 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g"] Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.450713 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.451597 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.451819 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.460634 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.555380 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-proxy-ca-bundles\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.555436 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-client-ca\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.555468 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-config\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.555511 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8mm8\" (UniqueName: \"kubernetes.io/projected/3a313caf-d8d9-4baf-992a-2682ba85bc34-kube-api-access-c8mm8\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.555561 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a313caf-d8d9-4baf-992a-2682ba85bc34-serving-cert\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.656808 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-client-ca\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.656928 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-config\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.656996 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8mm8\" (UniqueName: \"kubernetes.io/projected/3a313caf-d8d9-4baf-992a-2682ba85bc34-kube-api-access-c8mm8\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.657075 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a313caf-d8d9-4baf-992a-2682ba85bc34-serving-cert\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.657183 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-proxy-ca-bundles\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.658567 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-client-ca\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.658727 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-config\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.662312 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-proxy-ca-bundles\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.667610 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a313caf-d8d9-4baf-992a-2682ba85bc34-serving-cert\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.677705 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8mm8\" (UniqueName: \"kubernetes.io/projected/3a313caf-d8d9-4baf-992a-2682ba85bc34-kube-api-access-c8mm8\") pod \"controller-manager-7c8fc586d6-k9g2g\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.686098 4767 generic.go:334] "Generic (PLEG): container finished" podID="83954719-42d1-429e-aada-b80c003547bc" containerID="19983901980d0b9c7e6b5c4c0662bbe9a1d79b4d8aba2963aeba3fa5af32fa06" exitCode=0 Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.686152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xps7s" event={"ID":"83954719-42d1-429e-aada-b80c003547bc","Type":"ContainerDied","Data":"19983901980d0b9c7e6b5c4c0662bbe9a1d79b4d8aba2963aeba3fa5af32fa06"} Jan 29 15:05:03 crc kubenswrapper[4767]: I0129 15:05:03.768626 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.184827 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g"] Jan 29 15:05:04 crc kubenswrapper[4767]: W0129 15:05:04.201867 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a313caf_d8d9_4baf_992a_2682ba85bc34.slice/crio-1bc79916b867244fc1b3d727581addde57d6fa1945ea5a84dab52da15d7b6288 WatchSource:0}: Error finding container 1bc79916b867244fc1b3d727581addde57d6fa1945ea5a84dab52da15d7b6288: Status 404 returned error can't find the container with id 1bc79916b867244fc1b3d727581addde57d6fa1945ea5a84dab52da15d7b6288 Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.691987 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xps7s" event={"ID":"83954719-42d1-429e-aada-b80c003547bc","Type":"ContainerStarted","Data":"3b426efe2e1e496f7588512f79ed555f415f89dbff7f6f0a9d4ba70909bf2317"} Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.694423 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" event={"ID":"3a313caf-d8d9-4baf-992a-2682ba85bc34","Type":"ContainerStarted","Data":"be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d"} Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.694468 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" event={"ID":"3a313caf-d8d9-4baf-992a-2682ba85bc34","Type":"ContainerStarted","Data":"1bc79916b867244fc1b3d727581addde57d6fa1945ea5a84dab52da15d7b6288"} Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.694618 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.698938 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.711914 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xps7s" podStartSLOduration=3.48834434 podStartE2EDuration="1m37.71187782s" podCreationTimestamp="2026-01-29 15:03:27 +0000 UTC" firstStartedPulling="2026-01-29 15:03:29.929200532 +0000 UTC m=+165.739517571" lastFinishedPulling="2026-01-29 15:05:04.152733972 +0000 UTC m=+259.963051051" observedRunningTime="2026-01-29 15:05:04.711078176 +0000 UTC m=+260.521395215" watchObservedRunningTime="2026-01-29 15:05:04.71187782 +0000 UTC m=+260.522194859" Jan 29 15:05:04 crc kubenswrapper[4767]: I0129 15:05:04.734761 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" podStartSLOduration=5.734741072 podStartE2EDuration="5.734741072s" podCreationTimestamp="2026-01-29 15:04:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:05:04.729546247 +0000 UTC m=+260.539863286" watchObservedRunningTime="2026-01-29 15:05:04.734741072 +0000 UTC m=+260.545058121" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.187227 4767 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.189239 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.189530 4767 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.189942 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b" gracePeriod=15 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.189983 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44" gracePeriod=15 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190010 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc" gracePeriod=15 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190053 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1" gracePeriod=15 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190505 4767 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.190758 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190772 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.190789 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190798 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.190808 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190815 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.190828 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190834 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.190841 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190848 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.190860 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190866 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.190875 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190883 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191008 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191021 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191034 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191044 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.190997 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c" gracePeriod=15 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191052 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191180 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.191460 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191489 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.191673 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.239041 4767 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278319 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278372 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278416 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278463 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278492 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278517 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278708 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.278754 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.379759 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.379831 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.379920 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.379926 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380004 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380016 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380011 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380097 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380118 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380159 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380197 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380303 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380326 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380353 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.380396 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.540045 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:05 crc kubenswrapper[4767]: W0129 15:05:05.562961 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-cfbbbd7e165a87cc535c57e3383f1dbffbb895867e492f6b46d71a904c94937f WatchSource:0}: Error finding container cfbbbd7e165a87cc535c57e3383f1dbffbb895867e492f6b46d71a904c94937f: Status 404 returned error can't find the container with id cfbbbd7e165a87cc535c57e3383f1dbffbb895867e492f6b46d71a904c94937f Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.569534 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188f3bf5d2b7339a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 15:05:05.568142234 +0000 UTC m=+261.378459283,LastTimestamp:2026-01-29 15:05:05.568142234 +0000 UTC m=+261.378459283,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.701974 4767 generic.go:334] "Generic (PLEG): container finished" podID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" containerID="a4e268ffd57fc7adb97acd6731daf2b1e69cf294b70684f4415f5b1b7592f154" exitCode=0 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.702057 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f","Type":"ContainerDied","Data":"a4e268ffd57fc7adb97acd6731daf2b1e69cf294b70684f4415f5b1b7592f154"} Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.702819 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.703234 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.704062 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cfbbbd7e165a87cc535c57e3383f1dbffbb895867e492f6b46d71a904c94937f"} Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.707570 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.709880 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.710888 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c" exitCode=0 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.710934 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc" exitCode=0 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.710938 4767 scope.go:117] "RemoveContainer" containerID="4dfd76fc3fbf6ba6721742e660a2c2890e3e38784596e305560ddb0c044f53c9" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.710945 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44" exitCode=0 Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.711036 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1" exitCode=2 Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.769426 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.769955 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.770317 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.770617 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.770920 4767 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:05 crc kubenswrapper[4767]: I0129 15:05:05.770958 4767 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.771206 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="200ms" Jan 29 15:05:05 crc kubenswrapper[4767]: E0129 15:05:05.972812 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="400ms" Jan 29 15:05:06 crc kubenswrapper[4767]: E0129 15:05:06.374025 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="800ms" Jan 29 15:05:06 crc kubenswrapper[4767]: I0129 15:05:06.719999 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 15:05:06 crc kubenswrapper[4767]: I0129 15:05:06.722513 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67"} Jan 29 15:05:06 crc kubenswrapper[4767]: I0129 15:05:06.723745 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:06 crc kubenswrapper[4767]: E0129 15:05:06.725541 4767 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:06 crc kubenswrapper[4767]: I0129 15:05:06.725571 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.003758 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.004385 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.004871 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.102584 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kube-api-access\") pod \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.102720 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kubelet-dir\") pod \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.102757 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-var-lock\") pod \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\" (UID: \"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f\") " Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.102941 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" (UID: "38ef7a4d-41ef-4e64-8aa2-351c6ab2951f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.102965 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-var-lock" (OuterVolumeSpecName: "var-lock") pod "38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" (UID: "38ef7a4d-41ef-4e64-8aa2-351c6ab2951f"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.106802 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" (UID: "38ef7a4d-41ef-4e64-8aa2-351c6ab2951f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.175377 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="1.6s" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.204124 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.204164 4767 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.204173 4767 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/38ef7a4d-41ef-4e64-8aa2-351c6ab2951f-var-lock\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.579122 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.581698 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.582488 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.582855 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.713742 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.713817 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.713850 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.713862 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.713916 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.714055 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.714076 4767 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.714151 4767 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.728120 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"38ef7a4d-41ef-4e64-8aa2-351c6ab2951f","Type":"ContainerDied","Data":"514ca9e58ac272faf454ddf2639c43a81ca40491183f38be85639ff50ca631d1"} Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.728161 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="514ca9e58ac272faf454ddf2639c43a81ca40491183f38be85639ff50ca631d1" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.728161 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.731490 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.732353 4767 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b" exitCode=0 Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.732500 4767 scope.go:117] "RemoveContainer" containerID="1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.732666 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.733085 4767 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.741911 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.742481 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.746466 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.746630 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.752864 4767 scope.go:117] "RemoveContainer" containerID="38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.766945 4767 scope.go:117] "RemoveContainer" containerID="96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.780419 4767 scope.go:117] "RemoveContainer" containerID="7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.792644 4767 scope.go:117] "RemoveContainer" containerID="4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.808562 4767 scope.go:117] "RemoveContainer" containerID="05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.815812 4767 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.828026 4767 scope.go:117] "RemoveContainer" containerID="1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.830250 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\": container with ID starting with 1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c not found: ID does not exist" containerID="1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.830289 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c"} err="failed to get container status \"1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\": rpc error: code = NotFound desc = could not find container \"1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c\": container with ID starting with 1c39554250c4c093a768dda35033fe7b8577bbd3a0bf2d1c3a8984741c3f3c1c not found: ID does not exist" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.830322 4767 scope.go:117] "RemoveContainer" containerID="38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.830522 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\": container with ID starting with 38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc not found: ID does not exist" containerID="38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.830551 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc"} err="failed to get container status \"38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\": rpc error: code = NotFound desc = could not find container \"38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc\": container with ID starting with 38a59d08231acc335b29b3f052c2d88d7af4a8f898192317655d37deb4a7dbfc not found: ID does not exist" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.830570 4767 scope.go:117] "RemoveContainer" containerID="96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.830982 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\": container with ID starting with 96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44 not found: ID does not exist" containerID="96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.831006 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44"} err="failed to get container status \"96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\": rpc error: code = NotFound desc = could not find container \"96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44\": container with ID starting with 96ec48a2f5be23ddad358a77dc8d414a485368a04aeaed867acbb22b88143e44 not found: ID does not exist" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.831023 4767 scope.go:117] "RemoveContainer" containerID="7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.831290 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\": container with ID starting with 7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1 not found: ID does not exist" containerID="7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.831314 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1"} err="failed to get container status \"7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\": rpc error: code = NotFound desc = could not find container \"7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1\": container with ID starting with 7302a5534bd5014b1b7c48915909b0705d5585067c5b99ba7e9bbb4d4b4b29a1 not found: ID does not exist" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.831330 4767 scope.go:117] "RemoveContainer" containerID="4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.831513 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\": container with ID starting with 4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b not found: ID does not exist" containerID="4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.831535 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b"} err="failed to get container status \"4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\": rpc error: code = NotFound desc = could not find container \"4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b\": container with ID starting with 4a979bfb7b64cbb6d9fd610cff70a620cfa7b051db2e5111cacb8630ace3b39b not found: ID does not exist" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.831551 4767 scope.go:117] "RemoveContainer" containerID="05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0" Jan 29 15:05:07 crc kubenswrapper[4767]: E0129 15:05:07.831774 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\": container with ID starting with 05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0 not found: ID does not exist" containerID="05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0" Jan 29 15:05:07 crc kubenswrapper[4767]: I0129 15:05:07.831827 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0"} err="failed to get container status \"05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\": rpc error: code = NotFound desc = could not find container \"05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0\": container with ID starting with 05f17cfbd1119ef8b4c3147774109a99ce5a459a55b8cee6f8309f3cb80f62f0 not found: ID does not exist" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.294548 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" containerName="oauth-openshift" containerID="cri-o://6e2a29e8588d3839007b5c69acf683073b4f741648b866403a47a763e3cd9d62" gracePeriod=15 Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.297366 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.297406 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.746338 4767 generic.go:334] "Generic (PLEG): container finished" podID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" containerID="6e2a29e8588d3839007b5c69acf683073b4f741648b866403a47a763e3cd9d62" exitCode=0 Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.746446 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" event={"ID":"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3","Type":"ContainerDied","Data":"6e2a29e8588d3839007b5c69acf683073b4f741648b866403a47a763e3cd9d62"} Jan 29 15:05:08 crc kubenswrapper[4767]: E0129 15:05:08.777179 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="3.2s" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.820403 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.820851 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.821184 4767 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.821649 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930325 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-dir\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930374 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-router-certs\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930406 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mcrc\" (UniqueName: \"kubernetes.io/projected/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-kube-api-access-2mcrc\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930433 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-service-ca\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930464 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-trusted-ca-bundle\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930462 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930501 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-cliconfig\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930521 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-session\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930541 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-login\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930568 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-provider-selection\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930600 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-idp-0-file-data\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930626 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-error\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930641 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-ocp-branding-template\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930660 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-policies\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930679 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-serving-cert\") pod \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\" (UID: \"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3\") " Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.930868 4767 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.931153 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.931933 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.932241 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.932488 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.936124 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.936170 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-kube-api-access-2mcrc" (OuterVolumeSpecName: "kube-api-access-2mcrc") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "kube-api-access-2mcrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.936864 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.945071 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.946944 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.953094 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.953467 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.953559 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:08 crc kubenswrapper[4767]: I0129 15:05:08.958638 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" (UID: "b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.027692 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031667 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031698 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031709 4767 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031719 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031729 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031739 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mcrc\" (UniqueName: \"kubernetes.io/projected/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-kube-api-access-2mcrc\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031747 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031757 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031766 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031776 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031785 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031796 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.031805 4767 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.340807 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xps7s" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="registry-server" probeResult="failure" output=< Jan 29 15:05:09 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:05:09 crc kubenswrapper[4767]: > Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.752569 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" event={"ID":"b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3","Type":"ContainerDied","Data":"53af6631fe1ff4f467a0eb3653399d4e112388f97c6ab6f1f01157e5650781ba"} Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.752629 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.753236 4767 scope.go:117] "RemoveContainer" containerID="6e2a29e8588d3839007b5c69acf683073b4f741648b866403a47a763e3cd9d62" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.754034 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.754458 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.757237 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:09 crc kubenswrapper[4767]: I0129 15:05:09.757853 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:10 crc kubenswrapper[4767]: I0129 15:05:10.018924 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:10 crc kubenswrapper[4767]: I0129 15:05:10.020122 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:10 crc kubenswrapper[4767]: I0129 15:05:10.020487 4767 status_manager.go:851] "Failed to get status for pod" podUID="46581033-eca1-4394-9760-cf70446cb3ca" pod="openshift-marketplace/certified-operators-x5gw6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-x5gw6\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:11 crc kubenswrapper[4767]: I0129 15:05:11.774862 4767 generic.go:334] "Generic (PLEG): container finished" podID="46581033-eca1-4394-9760-cf70446cb3ca" containerID="c88872b91734d6249096c2079641a99b944d48d05cbd7dd5a54a1d8f619eb03c" exitCode=0 Jan 29 15:05:11 crc kubenswrapper[4767]: I0129 15:05:11.775389 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5gw6" event={"ID":"46581033-eca1-4394-9760-cf70446cb3ca","Type":"ContainerDied","Data":"c88872b91734d6249096c2079641a99b944d48d05cbd7dd5a54a1d8f619eb03c"} Jan 29 15:05:11 crc kubenswrapper[4767]: I0129 15:05:11.778029 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:11 crc kubenswrapper[4767]: I0129 15:05:11.778253 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:11 crc kubenswrapper[4767]: I0129 15:05:11.778471 4767 status_manager.go:851] "Failed to get status for pod" podUID="46581033-eca1-4394-9760-cf70446cb3ca" pod="openshift-marketplace/certified-operators-x5gw6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-x5gw6\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:11 crc kubenswrapper[4767]: E0129 15:05:11.979096 4767 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.138:6443: connect: connection refused" interval="6.4s" Jan 29 15:05:13 crc kubenswrapper[4767]: E0129 15:05:13.287218 4767 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.138:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188f3bf5d2b7339a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-29 15:05:05.568142234 +0000 UTC m=+261.378459283,LastTimestamp:2026-01-29 15:05:05.568142234 +0000 UTC m=+261.378459283,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 29 15:05:13 crc kubenswrapper[4767]: I0129 15:05:13.792764 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5gw6" event={"ID":"46581033-eca1-4394-9760-cf70446cb3ca","Type":"ContainerStarted","Data":"2c4e20e57f307267abf91f101db1e3d7dddd3c73e77068ff2bbfd586f053a7b2"} Jan 29 15:05:13 crc kubenswrapper[4767]: I0129 15:05:13.793732 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:13 crc kubenswrapper[4767]: I0129 15:05:13.794309 4767 status_manager.go:851] "Failed to get status for pod" podUID="46581033-eca1-4394-9760-cf70446cb3ca" pod="openshift-marketplace/certified-operators-x5gw6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-x5gw6\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:13 crc kubenswrapper[4767]: I0129 15:05:13.794832 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:15 crc kubenswrapper[4767]: I0129 15:05:15.021720 4767 status_manager.go:851] "Failed to get status for pod" podUID="46581033-eca1-4394-9760-cf70446cb3ca" pod="openshift-marketplace/certified-operators-x5gw6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-x5gw6\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:15 crc kubenswrapper[4767]: I0129 15:05:15.022453 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:15 crc kubenswrapper[4767]: I0129 15:05:15.022758 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:15 crc kubenswrapper[4767]: E0129 15:05:15.060173 4767 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" volumeName="registry-storage" Jan 29 15:05:15 crc kubenswrapper[4767]: I0129 15:05:15.097413 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:05:15 crc kubenswrapper[4767]: I0129 15:05:15.097454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.017972 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.019193 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.019890 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.020712 4767 status_manager.go:851] "Failed to get status for pod" podUID="46581033-eca1-4394-9760-cf70446cb3ca" pod="openshift-marketplace/certified-operators-x5gw6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-x5gw6\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.042522 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.043208 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:16 crc kubenswrapper[4767]: E0129 15:05:16.044076 4767 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.044855 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:16 crc kubenswrapper[4767]: W0129 15:05:16.081489 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-a0968b2d22079726490761c38e82cd9e1c09d41d7cf61821ba8110a6c594aa64 WatchSource:0}: Error finding container a0968b2d22079726490761c38e82cd9e1c09d41d7cf61821ba8110a6c594aa64: Status 404 returned error can't find the container with id a0968b2d22079726490761c38e82cd9e1c09d41d7cf61821ba8110a6c594aa64 Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.161112 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-x5gw6" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="registry-server" probeResult="failure" output=< Jan 29 15:05:16 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:05:16 crc kubenswrapper[4767]: > Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.812581 4767 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="da7bc81db50278d8a19b0c6985590711a54810dadc6ea65de7b61a4e17727b34" exitCode=0 Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.812669 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"da7bc81db50278d8a19b0c6985590711a54810dadc6ea65de7b61a4e17727b34"} Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.812897 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a0968b2d22079726490761c38e82cd9e1c09d41d7cf61821ba8110a6c594aa64"} Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.813270 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.813290 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.813563 4767 status_manager.go:851] "Failed to get status for pod" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:16 crc kubenswrapper[4767]: E0129 15:05:16.813567 4767 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.138:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.813885 4767 status_manager.go:851] "Failed to get status for pod" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" pod="openshift-authentication/oauth-openshift-558db77b4-4xbzf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4xbzf\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:16 crc kubenswrapper[4767]: I0129 15:05:16.814175 4767 status_manager.go:851] "Failed to get status for pod" podUID="46581033-eca1-4394-9760-cf70446cb3ca" pod="openshift-marketplace/certified-operators-x5gw6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-x5gw6\": dial tcp 38.102.83.138:6443: connect: connection refused" Jan 29 15:05:17 crc kubenswrapper[4767]: I0129 15:05:17.827541 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0e5d0bf6506b260ef94a200e5ba3ce795ff34b025919abd882bb0f66f04cbd13"} Jan 29 15:05:17 crc kubenswrapper[4767]: I0129 15:05:17.828090 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e07c72404e31eba30dbb3fcce382c2978b1afcd2a33e76916be7ab842b79b304"} Jan 29 15:05:17 crc kubenswrapper[4767]: I0129 15:05:17.828103 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ecaca9fff7457a6c59f2c2a3de130438e83d642202b48bf149b4a591727eb20a"} Jan 29 15:05:17 crc kubenswrapper[4767]: I0129 15:05:17.828112 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f3f67fe8ecfbb0b13dafe58a81e8765e1b686badb99eb8141fd08ec18d1ad16a"} Jan 29 15:05:18 crc kubenswrapper[4767]: I0129 15:05:18.340979 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:05:18 crc kubenswrapper[4767]: I0129 15:05:18.387712 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:05:18 crc kubenswrapper[4767]: I0129 15:05:18.834874 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a25246accab1ab765788d94e110a4c02782e36ba6c981abae596f837fdfd77dc"} Jan 29 15:05:18 crc kubenswrapper[4767]: I0129 15:05:18.835113 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:18 crc kubenswrapper[4767]: I0129 15:05:18.835126 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:19 crc kubenswrapper[4767]: I0129 15:05:19.843703 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 15:05:19 crc kubenswrapper[4767]: I0129 15:05:19.844016 4767 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace" exitCode=1 Jan 29 15:05:19 crc kubenswrapper[4767]: I0129 15:05:19.844052 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace"} Jan 29 15:05:19 crc kubenswrapper[4767]: I0129 15:05:19.844549 4767 scope.go:117] "RemoveContainer" containerID="c2852bfa4525a7c9ac00e6a10d1050628b926444c1ffa79c7c2ff23a7a717ace" Jan 29 15:05:20 crc kubenswrapper[4767]: I0129 15:05:20.305351 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:05:20 crc kubenswrapper[4767]: I0129 15:05:20.854192 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 29 15:05:20 crc kubenswrapper[4767]: I0129 15:05:20.854487 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ba05814f6070ab2a02672a719e4de9e8f1e7e040a4c3981454cf72becba8a6b5"} Jan 29 15:05:21 crc kubenswrapper[4767]: I0129 15:05:21.045200 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:21 crc kubenswrapper[4767]: I0129 15:05:21.045249 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:21 crc kubenswrapper[4767]: I0129 15:05:21.051084 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:21 crc kubenswrapper[4767]: I0129 15:05:21.588023 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:05:23 crc kubenswrapper[4767]: I0129 15:05:23.845225 4767 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:23 crc kubenswrapper[4767]: I0129 15:05:23.872165 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:23 crc kubenswrapper[4767]: I0129 15:05:23.872320 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:23 crc kubenswrapper[4767]: I0129 15:05:23.872353 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:23 crc kubenswrapper[4767]: I0129 15:05:23.878402 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:24 crc kubenswrapper[4767]: I0129 15:05:24.879437 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:24 crc kubenswrapper[4767]: I0129 15:05:24.879494 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:25 crc kubenswrapper[4767]: I0129 15:05:25.047963 4767 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5b5e7e11-e5ea-4b24-8b26-ce04dd213033" Jan 29 15:05:25 crc kubenswrapper[4767]: I0129 15:05:25.151307 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:05:25 crc kubenswrapper[4767]: I0129 15:05:25.186254 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:05:25 crc kubenswrapper[4767]: I0129 15:05:25.883484 4767 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:25 crc kubenswrapper[4767]: I0129 15:05:25.883511 4767 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="45069754-41c6-428c-bef5-149c235f5944" Jan 29 15:05:25 crc kubenswrapper[4767]: I0129 15:05:25.886859 4767 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="5b5e7e11-e5ea-4b24-8b26-ce04dd213033" Jan 29 15:05:28 crc kubenswrapper[4767]: I0129 15:05:28.002255 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:05:28 crc kubenswrapper[4767]: I0129 15:05:28.006860 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:05:30 crc kubenswrapper[4767]: I0129 15:05:30.157697 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 29 15:05:30 crc kubenswrapper[4767]: I0129 15:05:30.549126 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 29 15:05:30 crc kubenswrapper[4767]: I0129 15:05:30.756848 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 29 15:05:30 crc kubenswrapper[4767]: I0129 15:05:30.952172 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.022071 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.162548 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.250877 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.265001 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.288991 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.400809 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.417969 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.592715 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.622692 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 29 15:05:31 crc kubenswrapper[4767]: I0129 15:05:31.935484 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 15:05:32 crc kubenswrapper[4767]: I0129 15:05:32.195951 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 15:05:32 crc kubenswrapper[4767]: I0129 15:05:32.430635 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 29 15:05:32 crc kubenswrapper[4767]: I0129 15:05:32.445987 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 29 15:05:32 crc kubenswrapper[4767]: I0129 15:05:32.568822 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 15:05:33 crc kubenswrapper[4767]: I0129 15:05:33.556175 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 29 15:05:34 crc kubenswrapper[4767]: I0129 15:05:34.244818 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 29 15:05:34 crc kubenswrapper[4767]: I0129 15:05:34.619781 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 29 15:05:34 crc kubenswrapper[4767]: I0129 15:05:34.781755 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 29 15:05:35 crc kubenswrapper[4767]: I0129 15:05:35.737163 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 29 15:05:35 crc kubenswrapper[4767]: I0129 15:05:35.823710 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.488710 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.649714 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.665828 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.691437 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.774017 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.838187 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.891433 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 29 15:05:36 crc kubenswrapper[4767]: I0129 15:05:36.955540 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.041949 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.079569 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.356442 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.565067 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.635695 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.763591 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.765752 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.771216 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.819398 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 15:05:37 crc kubenswrapper[4767]: I0129 15:05:37.989954 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.002554 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.449840 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.514476 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.593498 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.664735 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.672775 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.687968 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.735480 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.791282 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 29 15:05:38 crc kubenswrapper[4767]: I0129 15:05:38.857833 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.061935 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.101976 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.242595 4767 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.244440 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x5gw6" podStartSLOduration=29.271989741 podStartE2EDuration="2m15.244421354s" podCreationTimestamp="2026-01-29 15:03:24 +0000 UTC" firstStartedPulling="2026-01-29 15:03:26.838326457 +0000 UTC m=+162.648643496" lastFinishedPulling="2026-01-29 15:05:12.81075803 +0000 UTC m=+268.621075109" observedRunningTime="2026-01-29 15:05:23.829845985 +0000 UTC m=+279.640163024" watchObservedRunningTime="2026-01-29 15:05:39.244421354 +0000 UTC m=+295.054738393" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.247779 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-4xbzf"] Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.247854 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.252686 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.268197 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.268174616 podStartE2EDuration="16.268174616s" podCreationTimestamp="2026-01-29 15:05:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:05:39.266494266 +0000 UTC m=+295.076811325" watchObservedRunningTime="2026-01-29 15:05:39.268174616 +0000 UTC m=+295.078491675" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.281388 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.354181 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.472589 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.519062 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.532294 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.539137 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.539314 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.617448 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.657596 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.725160 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.786399 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.965404 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.974501 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 29 15:05:39 crc kubenswrapper[4767]: I0129 15:05:39.998787 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.061496 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.063181 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.365527 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.401161 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.443117 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.499604 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.522354 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.719555 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.719635 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.727756 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.738004 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.784611 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.878682 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.897148 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 29 15:05:40 crc kubenswrapper[4767]: I0129 15:05:40.999240 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.016329 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.024531 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" path="/var/lib/kubelet/pods/b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3/volumes" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.252000 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.255489 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.359726 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.405776 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.443561 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.552378 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.620814 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.730688 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.762696 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.822075 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.828689 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.878014 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 29 15:05:41 crc kubenswrapper[4767]: I0129 15:05:41.891206 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.009359 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.034080 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.083553 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.385090 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.439693 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.440598 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.445717 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.471069 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.471805 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.475354 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.706635 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.731933 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.847619 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 29 15:05:42 crc kubenswrapper[4767]: I0129 15:05:42.909608 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.038611 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.042784 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.055871 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.101495 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.107237 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.210859 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.272219 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.305963 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.434820 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.442804 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.450219 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.506682 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.545636 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.552627 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.553682 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.583812 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.606151 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.687366 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.752883 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.819007 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-84db4cfc58-9wbct"] Jan 29 15:05:43 crc kubenswrapper[4767]: E0129 15:05:43.819253 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" containerName="oauth-openshift" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.819269 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" containerName="oauth-openshift" Jan 29 15:05:43 crc kubenswrapper[4767]: E0129 15:05:43.819296 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" containerName="installer" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.819304 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" containerName="installer" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.819422 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b821b2b4-3a1a-45b3-9cc3-3c0de20bb7a3" containerName="oauth-openshift" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.819450 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ef7a4d-41ef-4e64-8aa2-351c6ab2951f" containerName="installer" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.819951 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.822770 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823056 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823211 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823266 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823056 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823218 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823456 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823552 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823729 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823749 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823787 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.823801 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.833659 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.842417 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-84db4cfc58-9wbct"] Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.849795 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.852537 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.908754 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eee33530-b164-47bf-a47a-bb81b38cf01f-audit-dir\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909032 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-error\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909149 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-service-ca\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909250 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kggxh\" (UniqueName: \"kubernetes.io/projected/eee33530-b164-47bf-a47a-bb81b38cf01f-kube-api-access-kggxh\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909346 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909432 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-router-certs\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909517 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909603 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-login\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909675 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909762 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909841 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.909937 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-audit-policies\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.910033 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.910171 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-session\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:43 crc kubenswrapper[4767]: I0129 15:05:43.939384 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.011555 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-session\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012001 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eee33530-b164-47bf-a47a-bb81b38cf01f-audit-dir\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012241 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-error\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012129 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eee33530-b164-47bf-a47a-bb81b38cf01f-audit-dir\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012512 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-service-ca\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012651 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kggxh\" (UniqueName: \"kubernetes.io/projected/eee33530-b164-47bf-a47a-bb81b38cf01f-kube-api-access-kggxh\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012693 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012714 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-router-certs\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012740 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012763 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-login\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012789 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012827 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012851 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012888 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-audit-policies\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.012933 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.016587 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-audit-policies\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.017258 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.017281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.017855 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-service-ca\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.019725 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-session\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.019836 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-login\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.021277 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.025520 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-error\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.025603 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.025683 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-router-certs\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.029040 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.030265 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eee33530-b164-47bf-a47a-bb81b38cf01f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.036609 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kggxh\" (UniqueName: \"kubernetes.io/projected/eee33530-b164-47bf-a47a-bb81b38cf01f-kube-api-access-kggxh\") pod \"oauth-openshift-84db4cfc58-9wbct\" (UID: \"eee33530-b164-47bf-a47a-bb81b38cf01f\") " pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.051862 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.128564 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.129274 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.142416 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.147083 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.156769 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.207007 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.301832 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.339004 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.385469 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.452110 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.464770 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.516178 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.561720 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.576022 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.592189 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.593335 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-84db4cfc58-9wbct"] Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.638235 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.649300 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.699456 4767 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.915368 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.990806 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" event={"ID":"eee33530-b164-47bf-a47a-bb81b38cf01f","Type":"ContainerStarted","Data":"f60806c8ded782b1503d3054749a05b9e68be65c6d9dec6f1abf4e6e926d1694"} Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.990847 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" event={"ID":"eee33530-b164-47bf-a47a-bb81b38cf01f","Type":"ContainerStarted","Data":"a78946880effc6488f7e84b3e8bee63a5c79c70e66fd3aa8b3dc3d6e3ea4aa54"} Jan 29 15:05:44 crc kubenswrapper[4767]: I0129 15:05:44.991219 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.022104 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" podStartSLOduration=62.02208926 podStartE2EDuration="1m2.02208926s" podCreationTimestamp="2026-01-29 15:04:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:05:45.018824203 +0000 UTC m=+300.829141242" watchObservedRunningTime="2026-01-29 15:05:45.02208926 +0000 UTC m=+300.832406299" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.034242 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.039312 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.228919 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.274132 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.278515 4767 patch_prober.go:28] interesting pod/oauth-openshift-84db4cfc58-9wbct container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.62:6443/healthz\": read tcp 10.217.0.2:49718->10.217.0.62:6443: read: connection reset by peer" start-of-body= Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.278560 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" podUID="eee33530-b164-47bf-a47a-bb81b38cf01f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.62:6443/healthz\": read tcp 10.217.0.2:49718->10.217.0.62:6443: read: connection reset by peer" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.339804 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.368268 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.597983 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.662248 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.683392 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.759267 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.997570 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-84db4cfc58-9wbct_eee33530-b164-47bf-a47a-bb81b38cf01f/oauth-openshift/0.log" Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.997618 4767 generic.go:334] "Generic (PLEG): container finished" podID="eee33530-b164-47bf-a47a-bb81b38cf01f" containerID="f60806c8ded782b1503d3054749a05b9e68be65c6d9dec6f1abf4e6e926d1694" exitCode=255 Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.997644 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" event={"ID":"eee33530-b164-47bf-a47a-bb81b38cf01f","Type":"ContainerDied","Data":"f60806c8ded782b1503d3054749a05b9e68be65c6d9dec6f1abf4e6e926d1694"} Jan 29 15:05:45 crc kubenswrapper[4767]: I0129 15:05:45.998554 4767 scope.go:117] "RemoveContainer" containerID="f60806c8ded782b1503d3054749a05b9e68be65c6d9dec6f1abf4e6e926d1694" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.033238 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.057168 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.096625 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.111229 4767 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.331430 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.356076 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.383549 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.391102 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.443762 4767 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.444128 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67" gracePeriod=5 Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.527265 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.530033 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.568793 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.704483 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.775023 4767 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.846854 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.873717 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.948920 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 29 15:05:46 crc kubenswrapper[4767]: I0129 15:05:46.959262 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.000830 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.005753 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-84db4cfc58-9wbct_eee33530-b164-47bf-a47a-bb81b38cf01f/oauth-openshift/1.log" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.006353 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-84db4cfc58-9wbct_eee33530-b164-47bf-a47a-bb81b38cf01f/oauth-openshift/0.log" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.006412 4767 generic.go:334] "Generic (PLEG): container finished" podID="eee33530-b164-47bf-a47a-bb81b38cf01f" containerID="eb856d85c29ce465dd09d24e9bbda5d28f392d4577908e0694de83aef7476c88" exitCode=255 Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.006445 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" event={"ID":"eee33530-b164-47bf-a47a-bb81b38cf01f","Type":"ContainerDied","Data":"eb856d85c29ce465dd09d24e9bbda5d28f392d4577908e0694de83aef7476c88"} Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.006489 4767 scope.go:117] "RemoveContainer" containerID="f60806c8ded782b1503d3054749a05b9e68be65c6d9dec6f1abf4e6e926d1694" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.007309 4767 scope.go:117] "RemoveContainer" containerID="eb856d85c29ce465dd09d24e9bbda5d28f392d4577908e0694de83aef7476c88" Jan 29 15:05:47 crc kubenswrapper[4767]: E0129 15:05:47.008163 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-84db4cfc58-9wbct_openshift-authentication(eee33530-b164-47bf-a47a-bb81b38cf01f)\"" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" podUID="eee33530-b164-47bf-a47a-bb81b38cf01f" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.156277 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.176758 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.218219 4767 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.256941 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.382423 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.453786 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.514369 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.545706 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.726242 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.830726 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.920643 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.932578 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 29 15:05:47 crc kubenswrapper[4767]: I0129 15:05:47.993536 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.014871 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-84db4cfc58-9wbct_eee33530-b164-47bf-a47a-bb81b38cf01f/oauth-openshift/1.log" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.015494 4767 scope.go:117] "RemoveContainer" containerID="eb856d85c29ce465dd09d24e9bbda5d28f392d4577908e0694de83aef7476c88" Jan 29 15:05:48 crc kubenswrapper[4767]: E0129 15:05:48.015683 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-84db4cfc58-9wbct_openshift-authentication(eee33530-b164-47bf-a47a-bb81b38cf01f)\"" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" podUID="eee33530-b164-47bf-a47a-bb81b38cf01f" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.173615 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.284050 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.285157 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.295329 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.311203 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.313482 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.441358 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.654212 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.677253 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.775807 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 29 15:05:48 crc kubenswrapper[4767]: I0129 15:05:48.792795 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.119330 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.142434 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.213436 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.428749 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.589358 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.704729 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.720382 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.849486 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.857653 4767 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.859285 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.859707 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.860185 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.886972 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 15:05:49 crc kubenswrapper[4767]: I0129 15:05:49.957075 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.080547 4767 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.171085 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.193957 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.194179 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.223030 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.245071 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.348023 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.363251 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.395946 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.400511 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.416927 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.612840 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.782217 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.821077 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 29 15:05:50 crc kubenswrapper[4767]: I0129 15:05:50.895312 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 29 15:05:51 crc kubenswrapper[4767]: I0129 15:05:51.091260 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 29 15:05:51 crc kubenswrapper[4767]: I0129 15:05:51.106083 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 29 15:05:51 crc kubenswrapper[4767]: I0129 15:05:51.165881 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 29 15:05:51 crc kubenswrapper[4767]: I0129 15:05:51.345792 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.030557 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.030859 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.037408 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.037483 4767 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67" exitCode=137 Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.037542 4767 scope.go:117] "RemoveContainer" containerID="a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.037557 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.050042 4767 scope.go:117] "RemoveContainer" containerID="a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67" Jan 29 15:05:52 crc kubenswrapper[4767]: E0129 15:05:52.050422 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67\": container with ID starting with a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67 not found: ID does not exist" containerID="a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.050454 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67"} err="failed to get container status \"a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67\": rpc error: code = NotFound desc = could not find container \"a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67\": container with ID starting with a332f93940d4441abce6faf4b6ba2265e441a3f207c2ed64d69b2f3d59d31a67 not found: ID does not exist" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.148792 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149269 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149296 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149392 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149460 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149602 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149626 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149725 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.149723 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.150728 4767 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.150751 4767 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.150764 4767 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.150776 4767 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.156558 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.251857 4767 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.347475 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 29 15:05:52 crc kubenswrapper[4767]: I0129 15:05:52.814171 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 29 15:05:53 crc kubenswrapper[4767]: I0129 15:05:53.024606 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 29 15:05:53 crc kubenswrapper[4767]: I0129 15:05:53.169677 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 29 15:05:54 crc kubenswrapper[4767]: I0129 15:05:54.147444 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:54 crc kubenswrapper[4767]: I0129 15:05:54.147509 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:05:54 crc kubenswrapper[4767]: I0129 15:05:54.148130 4767 scope.go:117] "RemoveContainer" containerID="eb856d85c29ce465dd09d24e9bbda5d28f392d4577908e0694de83aef7476c88" Jan 29 15:05:54 crc kubenswrapper[4767]: E0129 15:05:54.148341 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-84db4cfc58-9wbct_openshift-authentication(eee33530-b164-47bf-a47a-bb81b38cf01f)\"" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" podUID="eee33530-b164-47bf-a47a-bb81b38cf01f" Jan 29 15:05:59 crc kubenswrapper[4767]: I0129 15:05:59.985292 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g"] Jan 29 15:05:59 crc kubenswrapper[4767]: I0129 15:05:59.987058 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" podUID="3a313caf-d8d9-4baf-992a-2682ba85bc34" containerName="controller-manager" containerID="cri-o://be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d" gracePeriod=30 Jan 29 15:05:59 crc kubenswrapper[4767]: I0129 15:05:59.994088 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr"] Jan 29 15:05:59 crc kubenswrapper[4767]: I0129 15:05:59.994555 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" podUID="1d2e8f49-4924-4616-8ba7-ed0225358ace" containerName="route-controller-manager" containerID="cri-o://bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1" gracePeriod=30 Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.405216 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.410422 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.575784 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a313caf-d8d9-4baf-992a-2682ba85bc34-serving-cert\") pod \"3a313caf-d8d9-4baf-992a-2682ba85bc34\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.575865 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-client-ca\") pod \"3a313caf-d8d9-4baf-992a-2682ba85bc34\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.575955 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-config\") pod \"1d2e8f49-4924-4616-8ba7-ed0225358ace\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.576058 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmn9q\" (UniqueName: \"kubernetes.io/projected/1d2e8f49-4924-4616-8ba7-ed0225358ace-kube-api-access-rmn9q\") pod \"1d2e8f49-4924-4616-8ba7-ed0225358ace\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.576108 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-client-ca\") pod \"1d2e8f49-4924-4616-8ba7-ed0225358ace\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.576768 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-client-ca" (OuterVolumeSpecName: "client-ca") pod "3a313caf-d8d9-4baf-992a-2682ba85bc34" (UID: "3a313caf-d8d9-4baf-992a-2682ba85bc34"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.577014 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d2e8f49-4924-4616-8ba7-ed0225358ace-serving-cert\") pod \"1d2e8f49-4924-4616-8ba7-ed0225358ace\" (UID: \"1d2e8f49-4924-4616-8ba7-ed0225358ace\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.577059 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-proxy-ca-bundles\") pod \"3a313caf-d8d9-4baf-992a-2682ba85bc34\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.577151 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-config\") pod \"3a313caf-d8d9-4baf-992a-2682ba85bc34\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.577192 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8mm8\" (UniqueName: \"kubernetes.io/projected/3a313caf-d8d9-4baf-992a-2682ba85bc34-kube-api-access-c8mm8\") pod \"3a313caf-d8d9-4baf-992a-2682ba85bc34\" (UID: \"3a313caf-d8d9-4baf-992a-2682ba85bc34\") " Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.577554 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.578008 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-client-ca" (OuterVolumeSpecName: "client-ca") pod "1d2e8f49-4924-4616-8ba7-ed0225358ace" (UID: "1d2e8f49-4924-4616-8ba7-ed0225358ace"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.578226 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-config" (OuterVolumeSpecName: "config") pod "3a313caf-d8d9-4baf-992a-2682ba85bc34" (UID: "3a313caf-d8d9-4baf-992a-2682ba85bc34"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.578351 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3a313caf-d8d9-4baf-992a-2682ba85bc34" (UID: "3a313caf-d8d9-4baf-992a-2682ba85bc34"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.578403 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-config" (OuterVolumeSpecName: "config") pod "1d2e8f49-4924-4616-8ba7-ed0225358ace" (UID: "1d2e8f49-4924-4616-8ba7-ed0225358ace"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.582191 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a313caf-d8d9-4baf-992a-2682ba85bc34-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3a313caf-d8d9-4baf-992a-2682ba85bc34" (UID: "3a313caf-d8d9-4baf-992a-2682ba85bc34"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.583022 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d2e8f49-4924-4616-8ba7-ed0225358ace-kube-api-access-rmn9q" (OuterVolumeSpecName: "kube-api-access-rmn9q") pod "1d2e8f49-4924-4616-8ba7-ed0225358ace" (UID: "1d2e8f49-4924-4616-8ba7-ed0225358ace"). InnerVolumeSpecName "kube-api-access-rmn9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.583093 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a313caf-d8d9-4baf-992a-2682ba85bc34-kube-api-access-c8mm8" (OuterVolumeSpecName: "kube-api-access-c8mm8") pod "3a313caf-d8d9-4baf-992a-2682ba85bc34" (UID: "3a313caf-d8d9-4baf-992a-2682ba85bc34"). InnerVolumeSpecName "kube-api-access-c8mm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.583191 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d2e8f49-4924-4616-8ba7-ed0225358ace-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1d2e8f49-4924-4616-8ba7-ed0225358ace" (UID: "1d2e8f49-4924-4616-8ba7-ed0225358ace"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679562 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679626 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8mm8\" (UniqueName: \"kubernetes.io/projected/3a313caf-d8d9-4baf-992a-2682ba85bc34-kube-api-access-c8mm8\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679650 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a313caf-d8d9-4baf-992a-2682ba85bc34-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679672 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679693 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmn9q\" (UniqueName: \"kubernetes.io/projected/1d2e8f49-4924-4616-8ba7-ed0225358ace-kube-api-access-rmn9q\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679711 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d2e8f49-4924-4616-8ba7-ed0225358ace-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679729 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d2e8f49-4924-4616-8ba7-ed0225358ace-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:00 crc kubenswrapper[4767]: I0129 15:06:00.679746 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a313caf-d8d9-4baf-992a-2682ba85bc34-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.092336 4767 generic.go:334] "Generic (PLEG): container finished" podID="3a313caf-d8d9-4baf-992a-2682ba85bc34" containerID="be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d" exitCode=0 Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.092423 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.092492 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" event={"ID":"3a313caf-d8d9-4baf-992a-2682ba85bc34","Type":"ContainerDied","Data":"be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d"} Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.092549 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g" event={"ID":"3a313caf-d8d9-4baf-992a-2682ba85bc34","Type":"ContainerDied","Data":"1bc79916b867244fc1b3d727581addde57d6fa1945ea5a84dab52da15d7b6288"} Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.092583 4767 scope.go:117] "RemoveContainer" containerID="be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.096581 4767 generic.go:334] "Generic (PLEG): container finished" podID="1d2e8f49-4924-4616-8ba7-ed0225358ace" containerID="bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1" exitCode=0 Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.096642 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" event={"ID":"1d2e8f49-4924-4616-8ba7-ed0225358ace","Type":"ContainerDied","Data":"bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1"} Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.096686 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" event={"ID":"1d2e8f49-4924-4616-8ba7-ed0225358ace","Type":"ContainerDied","Data":"71a8f2bd3135fcbba97ad0c3b518b0ca235ce23b8afcb157b8c2467de950e564"} Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.096749 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.141238 4767 scope.go:117] "RemoveContainer" containerID="be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d" Jan 29 15:06:01 crc kubenswrapper[4767]: E0129 15:06:01.142074 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d\": container with ID starting with be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d not found: ID does not exist" containerID="be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.142160 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d"} err="failed to get container status \"be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d\": rpc error: code = NotFound desc = could not find container \"be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d\": container with ID starting with be8abbb8212f1c1f59a5125aedd8671337e78563857892e612c5628991112c4d not found: ID does not exist" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.142212 4767 scope.go:117] "RemoveContainer" containerID="bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.156760 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g"] Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.165700 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c8fc586d6-k9g2g"] Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.167105 4767 scope.go:117] "RemoveContainer" containerID="bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.171447 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr"] Jan 29 15:06:01 crc kubenswrapper[4767]: E0129 15:06:01.171722 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1\": container with ID starting with bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1 not found: ID does not exist" containerID="bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.171810 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1"} err="failed to get container status \"bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1\": rpc error: code = NotFound desc = could not find container \"bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1\": container with ID starting with bcfc4ad092ae897b3651d12baea14b6284bbbc4d13ed66b579894018b85984e1 not found: ID does not exist" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.175613 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-54f4c4644f-w9ckr"] Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.472528 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw"] Jan 29 15:06:01 crc kubenswrapper[4767]: E0129 15:06:01.473116 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a313caf-d8d9-4baf-992a-2682ba85bc34" containerName="controller-manager" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.473143 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a313caf-d8d9-4baf-992a-2682ba85bc34" containerName="controller-manager" Jan 29 15:06:01 crc kubenswrapper[4767]: E0129 15:06:01.473154 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.473161 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 15:06:01 crc kubenswrapper[4767]: E0129 15:06:01.473174 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d2e8f49-4924-4616-8ba7-ed0225358ace" containerName="route-controller-manager" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.473183 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d2e8f49-4924-4616-8ba7-ed0225358ace" containerName="route-controller-manager" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.473419 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a313caf-d8d9-4baf-992a-2682ba85bc34" containerName="controller-manager" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.473433 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.473443 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d2e8f49-4924-4616-8ba7-ed0225358ace" containerName="route-controller-manager" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.474776 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.475813 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-whk2m"] Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.476707 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.478550 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.478650 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.478558 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.480546 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.480547 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.480620 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.480801 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.480971 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.482714 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.484721 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.484980 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.486940 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.506056 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.510676 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-whk2m"] Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.527184 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw"] Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.592767 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-proxy-ca-bundles\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.592829 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5wz9\" (UniqueName: \"kubernetes.io/projected/17660e21-3a90-40c5-a7fd-30ec8ac27637-kube-api-access-j5wz9\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.592930 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-client-ca\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.592978 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-serving-cert\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.593025 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-config\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.593094 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17660e21-3a90-40c5-a7fd-30ec8ac27637-serving-cert\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.593155 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qgbv\" (UniqueName: \"kubernetes.io/projected/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-kube-api-access-8qgbv\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.593201 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-client-ca\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.593241 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-config\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694216 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qgbv\" (UniqueName: \"kubernetes.io/projected/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-kube-api-access-8qgbv\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694284 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-client-ca\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694314 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-config\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694374 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-proxy-ca-bundles\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694400 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5wz9\" (UniqueName: \"kubernetes.io/projected/17660e21-3a90-40c5-a7fd-30ec8ac27637-kube-api-access-j5wz9\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694423 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-client-ca\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694442 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-serving-cert\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694477 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-config\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.694520 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17660e21-3a90-40c5-a7fd-30ec8ac27637-serving-cert\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.697382 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-config\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.697690 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-client-ca\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.697958 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-client-ca\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.698263 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-config\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.705399 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17660e21-3a90-40c5-a7fd-30ec8ac27637-serving-cert\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.707535 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-proxy-ca-bundles\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.708251 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-serving-cert\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.730161 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qgbv\" (UniqueName: \"kubernetes.io/projected/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-kube-api-access-8qgbv\") pod \"controller-manager-66b4b547f8-whk2m\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.730474 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5wz9\" (UniqueName: \"kubernetes.io/projected/17660e21-3a90-40c5-a7fd-30ec8ac27637-kube-api-access-j5wz9\") pod \"route-controller-manager-85d64d5c8b-dm2gw\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.812085 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:01 crc kubenswrapper[4767]: I0129 15:06:01.832302 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:02 crc kubenswrapper[4767]: I0129 15:06:02.128851 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-whk2m"] Jan 29 15:06:02 crc kubenswrapper[4767]: I0129 15:06:02.281210 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw"] Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.025114 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d2e8f49-4924-4616-8ba7-ed0225358ace" path="/var/lib/kubelet/pods/1d2e8f49-4924-4616-8ba7-ed0225358ace/volumes" Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.026202 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a313caf-d8d9-4baf-992a-2682ba85bc34" path="/var/lib/kubelet/pods/3a313caf-d8d9-4baf-992a-2682ba85bc34/volumes" Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.114101 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" event={"ID":"17660e21-3a90-40c5-a7fd-30ec8ac27637","Type":"ContainerStarted","Data":"512d60939485b6cc0616dcecd6a999664c207230c40078dfcb6ff2aeb003b6af"} Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.114140 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" event={"ID":"17660e21-3a90-40c5-a7fd-30ec8ac27637","Type":"ContainerStarted","Data":"5f53fd8eaf38d6a2c391f075dcbf6a93dca1df2575c0f7b487839e30d5c5a556"} Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.114412 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.117477 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" event={"ID":"9529b0d9-ab21-4d97-9dcf-09818bd5bac2","Type":"ContainerStarted","Data":"d8bd6e74ff91e3465438f160b06caee57e67adfad9b85f79bdf7ed49c1869ebb"} Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.117525 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" event={"ID":"9529b0d9-ab21-4d97-9dcf-09818bd5bac2","Type":"ContainerStarted","Data":"fcccc3fb380918cf346cc1a193327fc9ac12486be72e3ebe9668f0eec1d95d74"} Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.117686 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.122116 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.124006 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.135281 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" podStartSLOduration=3.135257748 podStartE2EDuration="3.135257748s" podCreationTimestamp="2026-01-29 15:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:06:03.131885698 +0000 UTC m=+318.942202737" watchObservedRunningTime="2026-01-29 15:06:03.135257748 +0000 UTC m=+318.945574797" Jan 29 15:06:03 crc kubenswrapper[4767]: I0129 15:06:03.168121 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" podStartSLOduration=3.168096539 podStartE2EDuration="3.168096539s" podCreationTimestamp="2026-01-29 15:06:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:06:03.165707749 +0000 UTC m=+318.976024788" watchObservedRunningTime="2026-01-29 15:06:03.168096539 +0000 UTC m=+318.978413618" Jan 29 15:06:08 crc kubenswrapper[4767]: I0129 15:06:08.018999 4767 scope.go:117] "RemoveContainer" containerID="eb856d85c29ce465dd09d24e9bbda5d28f392d4577908e0694de83aef7476c88" Jan 29 15:06:09 crc kubenswrapper[4767]: I0129 15:06:09.164681 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-84db4cfc58-9wbct_eee33530-b164-47bf-a47a-bb81b38cf01f/oauth-openshift/1.log" Jan 29 15:06:09 crc kubenswrapper[4767]: I0129 15:06:09.165254 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" event={"ID":"eee33530-b164-47bf-a47a-bb81b38cf01f","Type":"ContainerStarted","Data":"d0ccb361a3e287a2495e9a352a05cb066a64003a4571c13300066e0645536a0b"} Jan 29 15:06:09 crc kubenswrapper[4767]: I0129 15:06:09.166032 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:06:09 crc kubenswrapper[4767]: I0129 15:06:09.177619 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-84db4cfc58-9wbct" Jan 29 15:06:19 crc kubenswrapper[4767]: I0129 15:06:19.957831 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-whk2m"] Jan 29 15:06:19 crc kubenswrapper[4767]: I0129 15:06:19.958513 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" podUID="9529b0d9-ab21-4d97-9dcf-09818bd5bac2" containerName="controller-manager" containerID="cri-o://d8bd6e74ff91e3465438f160b06caee57e67adfad9b85f79bdf7ed49c1869ebb" gracePeriod=30 Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.013637 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw"] Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.013875 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" podUID="17660e21-3a90-40c5-a7fd-30ec8ac27637" containerName="route-controller-manager" containerID="cri-o://512d60939485b6cc0616dcecd6a999664c207230c40078dfcb6ff2aeb003b6af" gracePeriod=30 Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.236790 4767 generic.go:334] "Generic (PLEG): container finished" podID="17660e21-3a90-40c5-a7fd-30ec8ac27637" containerID="512d60939485b6cc0616dcecd6a999664c207230c40078dfcb6ff2aeb003b6af" exitCode=0 Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.236881 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" event={"ID":"17660e21-3a90-40c5-a7fd-30ec8ac27637","Type":"ContainerDied","Data":"512d60939485b6cc0616dcecd6a999664c207230c40078dfcb6ff2aeb003b6af"} Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.242368 4767 generic.go:334] "Generic (PLEG): container finished" podID="9529b0d9-ab21-4d97-9dcf-09818bd5bac2" containerID="d8bd6e74ff91e3465438f160b06caee57e67adfad9b85f79bdf7ed49c1869ebb" exitCode=0 Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.242414 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" event={"ID":"9529b0d9-ab21-4d97-9dcf-09818bd5bac2","Type":"ContainerDied","Data":"d8bd6e74ff91e3465438f160b06caee57e67adfad9b85f79bdf7ed49c1869ebb"} Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.543632 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.593539 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.657886 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5wz9\" (UniqueName: \"kubernetes.io/projected/17660e21-3a90-40c5-a7fd-30ec8ac27637-kube-api-access-j5wz9\") pod \"17660e21-3a90-40c5-a7fd-30ec8ac27637\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.658100 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-client-ca\") pod \"17660e21-3a90-40c5-a7fd-30ec8ac27637\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.658128 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17660e21-3a90-40c5-a7fd-30ec8ac27637-serving-cert\") pod \"17660e21-3a90-40c5-a7fd-30ec8ac27637\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.658165 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-config\") pod \"17660e21-3a90-40c5-a7fd-30ec8ac27637\" (UID: \"17660e21-3a90-40c5-a7fd-30ec8ac27637\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.658770 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-client-ca" (OuterVolumeSpecName: "client-ca") pod "17660e21-3a90-40c5-a7fd-30ec8ac27637" (UID: "17660e21-3a90-40c5-a7fd-30ec8ac27637"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.659037 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-config" (OuterVolumeSpecName: "config") pod "17660e21-3a90-40c5-a7fd-30ec8ac27637" (UID: "17660e21-3a90-40c5-a7fd-30ec8ac27637"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.663211 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17660e21-3a90-40c5-a7fd-30ec8ac27637-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "17660e21-3a90-40c5-a7fd-30ec8ac27637" (UID: "17660e21-3a90-40c5-a7fd-30ec8ac27637"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.663336 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17660e21-3a90-40c5-a7fd-30ec8ac27637-kube-api-access-j5wz9" (OuterVolumeSpecName: "kube-api-access-j5wz9") pod "17660e21-3a90-40c5-a7fd-30ec8ac27637" (UID: "17660e21-3a90-40c5-a7fd-30ec8ac27637"). InnerVolumeSpecName "kube-api-access-j5wz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.759752 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-client-ca\") pod \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.759873 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-serving-cert\") pod \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.760038 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-config\") pod \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.760087 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qgbv\" (UniqueName: \"kubernetes.io/projected/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-kube-api-access-8qgbv\") pod \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.760121 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-proxy-ca-bundles\") pod \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\" (UID: \"9529b0d9-ab21-4d97-9dcf-09818bd5bac2\") " Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.760444 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.760465 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17660e21-3a90-40c5-a7fd-30ec8ac27637-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.760483 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17660e21-3a90-40c5-a7fd-30ec8ac27637-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.760500 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5wz9\" (UniqueName: \"kubernetes.io/projected/17660e21-3a90-40c5-a7fd-30ec8ac27637-kube-api-access-j5wz9\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.762022 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-client-ca" (OuterVolumeSpecName: "client-ca") pod "9529b0d9-ab21-4d97-9dcf-09818bd5bac2" (UID: "9529b0d9-ab21-4d97-9dcf-09818bd5bac2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.762153 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-config" (OuterVolumeSpecName: "config") pod "9529b0d9-ab21-4d97-9dcf-09818bd5bac2" (UID: "9529b0d9-ab21-4d97-9dcf-09818bd5bac2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.762175 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9529b0d9-ab21-4d97-9dcf-09818bd5bac2" (UID: "9529b0d9-ab21-4d97-9dcf-09818bd5bac2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.764463 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-kube-api-access-8qgbv" (OuterVolumeSpecName: "kube-api-access-8qgbv") pod "9529b0d9-ab21-4d97-9dcf-09818bd5bac2" (UID: "9529b0d9-ab21-4d97-9dcf-09818bd5bac2"). InnerVolumeSpecName "kube-api-access-8qgbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.764866 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9529b0d9-ab21-4d97-9dcf-09818bd5bac2" (UID: "9529b0d9-ab21-4d97-9dcf-09818bd5bac2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.862347 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.862385 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.862405 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.862423 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qgbv\" (UniqueName: \"kubernetes.io/projected/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-kube-api-access-8qgbv\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:20 crc kubenswrapper[4767]: I0129 15:06:20.862440 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9529b0d9-ab21-4d97-9dcf-09818bd5bac2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.256208 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" event={"ID":"9529b0d9-ab21-4d97-9dcf-09818bd5bac2","Type":"ContainerDied","Data":"fcccc3fb380918cf346cc1a193327fc9ac12486be72e3ebe9668f0eec1d95d74"} Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.256348 4767 scope.go:117] "RemoveContainer" containerID="d8bd6e74ff91e3465438f160b06caee57e67adfad9b85f79bdf7ed49c1869ebb" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.256674 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66b4b547f8-whk2m" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.262634 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" event={"ID":"17660e21-3a90-40c5-a7fd-30ec8ac27637","Type":"ContainerDied","Data":"5f53fd8eaf38d6a2c391f075dcbf6a93dca1df2575c0f7b487839e30d5c5a556"} Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.263082 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.291397 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw"] Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.291922 4767 scope.go:117] "RemoveContainer" containerID="512d60939485b6cc0616dcecd6a999664c207230c40078dfcb6ff2aeb003b6af" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.298626 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-dm2gw"] Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.303407 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-whk2m"] Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.308144 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-whk2m"] Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.487133 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-5w242"] Jan 29 15:06:21 crc kubenswrapper[4767]: E0129 15:06:21.487445 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17660e21-3a90-40c5-a7fd-30ec8ac27637" containerName="route-controller-manager" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.487468 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="17660e21-3a90-40c5-a7fd-30ec8ac27637" containerName="route-controller-manager" Jan 29 15:06:21 crc kubenswrapper[4767]: E0129 15:06:21.487484 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9529b0d9-ab21-4d97-9dcf-09818bd5bac2" containerName="controller-manager" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.487506 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="9529b0d9-ab21-4d97-9dcf-09818bd5bac2" containerName="controller-manager" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.487682 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="9529b0d9-ab21-4d97-9dcf-09818bd5bac2" containerName="controller-manager" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.487701 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="17660e21-3a90-40c5-a7fd-30ec8ac27637" containerName="route-controller-manager" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.488367 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.492627 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf"] Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.494498 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.495312 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.496051 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.496748 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.497541 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.498249 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.498350 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.499695 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.501310 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.502316 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-5w242"] Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.502386 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.507042 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf"] Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.507917 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.507936 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.507956 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.509103 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.677527 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gthz\" (UniqueName: \"kubernetes.io/projected/f9799634-f53e-406a-9918-135ba5f9d891-kube-api-access-7gthz\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.677615 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe74dbcd-7501-4f32-8e76-41dedc4a0955-serving-cert\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.677658 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-client-ca\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.677690 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-proxy-ca-bundles\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.677730 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-client-ca\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.677938 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-config\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.678189 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhp8h\" (UniqueName: \"kubernetes.io/projected/fe74dbcd-7501-4f32-8e76-41dedc4a0955-kube-api-access-jhp8h\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.678246 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9799634-f53e-406a-9918-135ba5f9d891-serving-cert\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.678279 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-config\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780117 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-config\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780613 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhp8h\" (UniqueName: \"kubernetes.io/projected/fe74dbcd-7501-4f32-8e76-41dedc4a0955-kube-api-access-jhp8h\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780665 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9799634-f53e-406a-9918-135ba5f9d891-serving-cert\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780700 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-config\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780747 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gthz\" (UniqueName: \"kubernetes.io/projected/f9799634-f53e-406a-9918-135ba5f9d891-kube-api-access-7gthz\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780794 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe74dbcd-7501-4f32-8e76-41dedc4a0955-serving-cert\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780831 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-client-ca\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780864 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-proxy-ca-bundles\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.780952 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-client-ca\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.782814 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-client-ca\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.784753 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-config\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.786156 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-client-ca\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.788748 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-config\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.789655 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-proxy-ca-bundles\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.793439 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe74dbcd-7501-4f32-8e76-41dedc4a0955-serving-cert\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.794885 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9799634-f53e-406a-9918-135ba5f9d891-serving-cert\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.807241 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhp8h\" (UniqueName: \"kubernetes.io/projected/fe74dbcd-7501-4f32-8e76-41dedc4a0955-kube-api-access-jhp8h\") pod \"route-controller-manager-75664bd6d9-2b9cf\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.817431 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gthz\" (UniqueName: \"kubernetes.io/projected/f9799634-f53e-406a-9918-135ba5f9d891-kube-api-access-7gthz\") pod \"controller-manager-658fd5994d-5w242\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:21 crc kubenswrapper[4767]: I0129 15:06:21.822482 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.052169 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf"] Jan 29 15:06:22 crc kubenswrapper[4767]: W0129 15:06:22.058648 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe74dbcd_7501_4f32_8e76_41dedc4a0955.slice/crio-3976cbeb0135f62af2228397683b60f924a501f386088fda4eb1756e6cb46e8e WatchSource:0}: Error finding container 3976cbeb0135f62af2228397683b60f924a501f386088fda4eb1756e6cb46e8e: Status 404 returned error can't find the container with id 3976cbeb0135f62af2228397683b60f924a501f386088fda4eb1756e6cb46e8e Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.112702 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.268563 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" event={"ID":"fe74dbcd-7501-4f32-8e76-41dedc4a0955","Type":"ContainerStarted","Data":"59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2"} Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.268640 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" event={"ID":"fe74dbcd-7501-4f32-8e76-41dedc4a0955","Type":"ContainerStarted","Data":"3976cbeb0135f62af2228397683b60f924a501f386088fda4eb1756e6cb46e8e"} Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.269692 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.271600 4767 patch_prober.go:28] interesting pod/route-controller-manager-75664bd6d9-2b9cf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.271648 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" podUID="fe74dbcd-7501-4f32-8e76-41dedc4a0955" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.289071 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" podStartSLOduration=2.289050656 podStartE2EDuration="2.289050656s" podCreationTimestamp="2026-01-29 15:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:06:22.287799779 +0000 UTC m=+338.098116818" watchObservedRunningTime="2026-01-29 15:06:22.289050656 +0000 UTC m=+338.099367695" Jan 29 15:06:22 crc kubenswrapper[4767]: I0129 15:06:22.513390 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-5w242"] Jan 29 15:06:22 crc kubenswrapper[4767]: W0129 15:06:22.519198 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9799634_f53e_406a_9918_135ba5f9d891.slice/crio-e942b8aee02409782932b0ca3336837d7b4b3d446648892af3437b406e257e85 WatchSource:0}: Error finding container e942b8aee02409782932b0ca3336837d7b4b3d446648892af3437b406e257e85: Status 404 returned error can't find the container with id e942b8aee02409782932b0ca3336837d7b4b3d446648892af3437b406e257e85 Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.024261 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17660e21-3a90-40c5-a7fd-30ec8ac27637" path="/var/lib/kubelet/pods/17660e21-3a90-40c5-a7fd-30ec8ac27637/volumes" Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.024880 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9529b0d9-ab21-4d97-9dcf-09818bd5bac2" path="/var/lib/kubelet/pods/9529b0d9-ab21-4d97-9dcf-09818bd5bac2/volumes" Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.279775 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" event={"ID":"f9799634-f53e-406a-9918-135ba5f9d891","Type":"ContainerStarted","Data":"aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0"} Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.279841 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" event={"ID":"f9799634-f53e-406a-9918-135ba5f9d891","Type":"ContainerStarted","Data":"e942b8aee02409782932b0ca3336837d7b4b3d446648892af3437b406e257e85"} Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.279909 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.288466 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.289326 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:06:23 crc kubenswrapper[4767]: I0129 15:06:23.309283 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" podStartSLOduration=4.30925523 podStartE2EDuration="4.30925523s" podCreationTimestamp="2026-01-29 15:06:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:06:23.306811547 +0000 UTC m=+339.117128616" watchObservedRunningTime="2026-01-29 15:06:23.30925523 +0000 UTC m=+339.119572309" Jan 29 15:06:30 crc kubenswrapper[4767]: I0129 15:06:30.158529 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5gw6"] Jan 29 15:06:30 crc kubenswrapper[4767]: I0129 15:06:30.161186 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x5gw6" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="registry-server" containerID="cri-o://2c4e20e57f307267abf91f101db1e3d7dddd3c73e77068ff2bbfd586f053a7b2" gracePeriod=2 Jan 29 15:06:31 crc kubenswrapper[4767]: I0129 15:06:31.322928 4767 generic.go:334] "Generic (PLEG): container finished" podID="46581033-eca1-4394-9760-cf70446cb3ca" containerID="2c4e20e57f307267abf91f101db1e3d7dddd3c73e77068ff2bbfd586f053a7b2" exitCode=0 Jan 29 15:06:31 crc kubenswrapper[4767]: I0129 15:06:31.322967 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5gw6" event={"ID":"46581033-eca1-4394-9760-cf70446cb3ca","Type":"ContainerDied","Data":"2c4e20e57f307267abf91f101db1e3d7dddd3c73e77068ff2bbfd586f053a7b2"} Jan 29 15:06:31 crc kubenswrapper[4767]: I0129 15:06:31.826717 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.013258 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-utilities\") pod \"46581033-eca1-4394-9760-cf70446cb3ca\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.013331 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grzjc\" (UniqueName: \"kubernetes.io/projected/46581033-eca1-4394-9760-cf70446cb3ca-kube-api-access-grzjc\") pod \"46581033-eca1-4394-9760-cf70446cb3ca\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.013379 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-catalog-content\") pod \"46581033-eca1-4394-9760-cf70446cb3ca\" (UID: \"46581033-eca1-4394-9760-cf70446cb3ca\") " Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.014504 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-utilities" (OuterVolumeSpecName: "utilities") pod "46581033-eca1-4394-9760-cf70446cb3ca" (UID: "46581033-eca1-4394-9760-cf70446cb3ca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.018974 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46581033-eca1-4394-9760-cf70446cb3ca-kube-api-access-grzjc" (OuterVolumeSpecName: "kube-api-access-grzjc") pod "46581033-eca1-4394-9760-cf70446cb3ca" (UID: "46581033-eca1-4394-9760-cf70446cb3ca"). InnerVolumeSpecName "kube-api-access-grzjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.066814 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46581033-eca1-4394-9760-cf70446cb3ca" (UID: "46581033-eca1-4394-9760-cf70446cb3ca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.114366 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grzjc\" (UniqueName: \"kubernetes.io/projected/46581033-eca1-4394-9760-cf70446cb3ca-kube-api-access-grzjc\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.114391 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.114401 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46581033-eca1-4394-9760-cf70446cb3ca-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.329541 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5gw6" event={"ID":"46581033-eca1-4394-9760-cf70446cb3ca","Type":"ContainerDied","Data":"aa7955a3fb9297e484db00e71cac6da77f576286e190fb6e9a2fabc256316171"} Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.329606 4767 scope.go:117] "RemoveContainer" containerID="2c4e20e57f307267abf91f101db1e3d7dddd3c73e77068ff2bbfd586f053a7b2" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.329795 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5gw6" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.349249 4767 scope.go:117] "RemoveContainer" containerID="c88872b91734d6249096c2079641a99b944d48d05cbd7dd5a54a1d8f619eb03c" Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.360953 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xps7s"] Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.361706 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xps7s" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="registry-server" containerID="cri-o://3b426efe2e1e496f7588512f79ed555f415f89dbff7f6f0a9d4ba70909bf2317" gracePeriod=2 Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.371672 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5gw6"] Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.376072 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x5gw6"] Jan 29 15:06:32 crc kubenswrapper[4767]: I0129 15:06:32.384988 4767 scope.go:117] "RemoveContainer" containerID="1c6456e08602fef5c153f8c5bcdc383169685d98ce5e2ecd260b9d942198f4b8" Jan 29 15:06:33 crc kubenswrapper[4767]: I0129 15:06:33.025391 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46581033-eca1-4394-9760-cf70446cb3ca" path="/var/lib/kubelet/pods/46581033-eca1-4394-9760-cf70446cb3ca/volumes" Jan 29 15:06:33 crc kubenswrapper[4767]: I0129 15:06:33.337241 4767 generic.go:334] "Generic (PLEG): container finished" podID="83954719-42d1-429e-aada-b80c003547bc" containerID="3b426efe2e1e496f7588512f79ed555f415f89dbff7f6f0a9d4ba70909bf2317" exitCode=0 Jan 29 15:06:33 crc kubenswrapper[4767]: I0129 15:06:33.337315 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xps7s" event={"ID":"83954719-42d1-429e-aada-b80c003547bc","Type":"ContainerDied","Data":"3b426efe2e1e496f7588512f79ed555f415f89dbff7f6f0a9d4ba70909bf2317"} Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.113699 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.240987 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-utilities\") pod \"83954719-42d1-429e-aada-b80c003547bc\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.241045 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29n4h\" (UniqueName: \"kubernetes.io/projected/83954719-42d1-429e-aada-b80c003547bc-kube-api-access-29n4h\") pod \"83954719-42d1-429e-aada-b80c003547bc\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.241074 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-catalog-content\") pod \"83954719-42d1-429e-aada-b80c003547bc\" (UID: \"83954719-42d1-429e-aada-b80c003547bc\") " Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.332588 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83954719-42d1-429e-aada-b80c003547bc-kube-api-access-29n4h" (OuterVolumeSpecName: "kube-api-access-29n4h") pod "83954719-42d1-429e-aada-b80c003547bc" (UID: "83954719-42d1-429e-aada-b80c003547bc"). InnerVolumeSpecName "kube-api-access-29n4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.343503 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29n4h\" (UniqueName: \"kubernetes.io/projected/83954719-42d1-429e-aada-b80c003547bc-kube-api-access-29n4h\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.347310 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xps7s" event={"ID":"83954719-42d1-429e-aada-b80c003547bc","Type":"ContainerDied","Data":"c0a4cfa95cbbee35ba59d470acb475ca92d4fa8d8f5d0e4711932e9f1f9cfe2b"} Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.347350 4767 scope.go:117] "RemoveContainer" containerID="3b426efe2e1e496f7588512f79ed555f415f89dbff7f6f0a9d4ba70909bf2317" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.347360 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xps7s" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.361245 4767 scope.go:117] "RemoveContainer" containerID="19983901980d0b9c7e6b5c4c0662bbe9a1d79b4d8aba2963aeba3fa5af32fa06" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.368101 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-utilities" (OuterVolumeSpecName: "utilities") pod "83954719-42d1-429e-aada-b80c003547bc" (UID: "83954719-42d1-429e-aada-b80c003547bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.375372 4767 scope.go:117] "RemoveContainer" containerID="17b8e4bbcd8cfb2b9dc8cfb69f90347f61b9c74bace2e8f72129e1553e4edf05" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.444203 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.514987 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83954719-42d1-429e-aada-b80c003547bc" (UID: "83954719-42d1-429e-aada-b80c003547bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.545884 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83954719-42d1-429e-aada-b80c003547bc-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.680188 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xps7s"] Jan 29 15:06:34 crc kubenswrapper[4767]: I0129 15:06:34.680238 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xps7s"] Jan 29 15:06:35 crc kubenswrapper[4767]: I0129 15:06:35.025847 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83954719-42d1-429e-aada-b80c003547bc" path="/var/lib/kubelet/pods/83954719-42d1-429e-aada-b80c003547bc/volumes" Jan 29 15:06:42 crc kubenswrapper[4767]: I0129 15:06:42.806015 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:06:42 crc kubenswrapper[4767]: I0129 15:06:42.806560 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:06:59 crc kubenswrapper[4767]: I0129 15:06:59.928477 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf"] Jan 29 15:06:59 crc kubenswrapper[4767]: I0129 15:06:59.929423 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" podUID="fe74dbcd-7501-4f32-8e76-41dedc4a0955" containerName="route-controller-manager" containerID="cri-o://59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2" gracePeriod=30 Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.344429 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.401877 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhp8h\" (UniqueName: \"kubernetes.io/projected/fe74dbcd-7501-4f32-8e76-41dedc4a0955-kube-api-access-jhp8h\") pod \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.401965 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-config\") pod \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.401998 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-client-ca\") pod \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.402049 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe74dbcd-7501-4f32-8e76-41dedc4a0955-serving-cert\") pod \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\" (UID: \"fe74dbcd-7501-4f32-8e76-41dedc4a0955\") " Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.403078 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-config" (OuterVolumeSpecName: "config") pod "fe74dbcd-7501-4f32-8e76-41dedc4a0955" (UID: "fe74dbcd-7501-4f32-8e76-41dedc4a0955"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.403782 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-client-ca" (OuterVolumeSpecName: "client-ca") pod "fe74dbcd-7501-4f32-8e76-41dedc4a0955" (UID: "fe74dbcd-7501-4f32-8e76-41dedc4a0955"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.409302 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe74dbcd-7501-4f32-8e76-41dedc4a0955-kube-api-access-jhp8h" (OuterVolumeSpecName: "kube-api-access-jhp8h") pod "fe74dbcd-7501-4f32-8e76-41dedc4a0955" (UID: "fe74dbcd-7501-4f32-8e76-41dedc4a0955"). InnerVolumeSpecName "kube-api-access-jhp8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.411932 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe74dbcd-7501-4f32-8e76-41dedc4a0955-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fe74dbcd-7501-4f32-8e76-41dedc4a0955" (UID: "fe74dbcd-7501-4f32-8e76-41dedc4a0955"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.482583 4767 generic.go:334] "Generic (PLEG): container finished" podID="fe74dbcd-7501-4f32-8e76-41dedc4a0955" containerID="59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2" exitCode=0 Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.482640 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" event={"ID":"fe74dbcd-7501-4f32-8e76-41dedc4a0955","Type":"ContainerDied","Data":"59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2"} Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.482667 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" event={"ID":"fe74dbcd-7501-4f32-8e76-41dedc4a0955","Type":"ContainerDied","Data":"3976cbeb0135f62af2228397683b60f924a501f386088fda4eb1756e6cb46e8e"} Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.482682 4767 scope.go:117] "RemoveContainer" containerID="59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.482827 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.503044 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhp8h\" (UniqueName: \"kubernetes.io/projected/fe74dbcd-7501-4f32-8e76-41dedc4a0955-kube-api-access-jhp8h\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.503080 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.503095 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe74dbcd-7501-4f32-8e76-41dedc4a0955-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.503106 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe74dbcd-7501-4f32-8e76-41dedc4a0955-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.504222 4767 scope.go:117] "RemoveContainer" containerID="59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2" Jan 29 15:07:00 crc kubenswrapper[4767]: E0129 15:07:00.504731 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2\": container with ID starting with 59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2 not found: ID does not exist" containerID="59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.504781 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2"} err="failed to get container status \"59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2\": rpc error: code = NotFound desc = could not find container \"59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2\": container with ID starting with 59ca73f34774e6be8269fa7c27afb2d3248c62f2d01161a898ce6ef88081f0f2 not found: ID does not exist" Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.514788 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf"] Jan 29 15:07:00 crc kubenswrapper[4767]: I0129 15:07:00.523129 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75664bd6d9-2b9cf"] Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.024484 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe74dbcd-7501-4f32-8e76-41dedc4a0955" path="/var/lib/kubelet/pods/fe74dbcd-7501-4f32-8e76-41dedc4a0955/volumes" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513205 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb"] Jan 29 15:07:01 crc kubenswrapper[4767]: E0129 15:07:01.513491 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="registry-server" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513508 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="registry-server" Jan 29 15:07:01 crc kubenswrapper[4767]: E0129 15:07:01.513525 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="extract-utilities" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513533 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="extract-utilities" Jan 29 15:07:01 crc kubenswrapper[4767]: E0129 15:07:01.513543 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe74dbcd-7501-4f32-8e76-41dedc4a0955" containerName="route-controller-manager" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513553 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe74dbcd-7501-4f32-8e76-41dedc4a0955" containerName="route-controller-manager" Jan 29 15:07:01 crc kubenswrapper[4767]: E0129 15:07:01.513575 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="extract-utilities" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513586 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="extract-utilities" Jan 29 15:07:01 crc kubenswrapper[4767]: E0129 15:07:01.513598 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="registry-server" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513608 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="registry-server" Jan 29 15:07:01 crc kubenswrapper[4767]: E0129 15:07:01.513625 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="extract-content" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513633 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="extract-content" Jan 29 15:07:01 crc kubenswrapper[4767]: E0129 15:07:01.513649 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="extract-content" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513657 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="extract-content" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513788 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="83954719-42d1-429e-aada-b80c003547bc" containerName="registry-server" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513810 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="46581033-eca1-4394-9760-cf70446cb3ca" containerName="registry-server" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.513832 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe74dbcd-7501-4f32-8e76-41dedc4a0955" containerName="route-controller-manager" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.514320 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.516771 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.518010 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.518014 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.518357 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.518519 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.518685 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.527072 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb"] Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.618374 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e463c882-7946-4780-8505-ec483466a35b-config\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.618709 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e463c882-7946-4780-8505-ec483466a35b-client-ca\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.618766 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e463c882-7946-4780-8505-ec483466a35b-serving-cert\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.618802 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjc69\" (UniqueName: \"kubernetes.io/projected/e463c882-7946-4780-8505-ec483466a35b-kube-api-access-cjc69\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.720314 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e463c882-7946-4780-8505-ec483466a35b-config\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.720388 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e463c882-7946-4780-8505-ec483466a35b-client-ca\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.720478 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e463c882-7946-4780-8505-ec483466a35b-serving-cert\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.720552 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjc69\" (UniqueName: \"kubernetes.io/projected/e463c882-7946-4780-8505-ec483466a35b-kube-api-access-cjc69\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.721980 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e463c882-7946-4780-8505-ec483466a35b-config\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.722074 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e463c882-7946-4780-8505-ec483466a35b-client-ca\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.730835 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e463c882-7946-4780-8505-ec483466a35b-serving-cert\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.747635 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjc69\" (UniqueName: \"kubernetes.io/projected/e463c882-7946-4780-8505-ec483466a35b-kube-api-access-cjc69\") pod \"route-controller-manager-85d64d5c8b-bsrxb\" (UID: \"e463c882-7946-4780-8505-ec483466a35b\") " pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:01 crc kubenswrapper[4767]: I0129 15:07:01.827716 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:02 crc kubenswrapper[4767]: I0129 15:07:02.297867 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb"] Jan 29 15:07:02 crc kubenswrapper[4767]: I0129 15:07:02.496507 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" event={"ID":"e463c882-7946-4780-8505-ec483466a35b","Type":"ContainerStarted","Data":"cbdc33714732ea800de28d71af299daf0bd31ba9f889224cba3bac7908bb2dc0"} Jan 29 15:07:02 crc kubenswrapper[4767]: I0129 15:07:02.496826 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:02 crc kubenswrapper[4767]: I0129 15:07:02.496841 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" event={"ID":"e463c882-7946-4780-8505-ec483466a35b","Type":"ContainerStarted","Data":"67b745706188a41c596615360abe62c092f66b9dcf5ed4c7759e840812feddd2"} Jan 29 15:07:02 crc kubenswrapper[4767]: I0129 15:07:02.511613 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" podStartSLOduration=3.5115966910000003 podStartE2EDuration="3.511596691s" podCreationTimestamp="2026-01-29 15:06:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:07:02.509133557 +0000 UTC m=+378.319450596" watchObservedRunningTime="2026-01-29 15:07:02.511596691 +0000 UTC m=+378.321913740" Jan 29 15:07:02 crc kubenswrapper[4767]: I0129 15:07:02.902172 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-85d64d5c8b-bsrxb" Jan 29 15:07:12 crc kubenswrapper[4767]: I0129 15:07:12.805710 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:07:12 crc kubenswrapper[4767]: I0129 15:07:12.806337 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.943663 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4fjjw"] Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.947587 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4fjjw" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="registry-server" containerID="cri-o://caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650" gracePeriod=30 Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.951977 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s7srb"] Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.952260 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s7srb" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="registry-server" containerID="cri-o://249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc" gracePeriod=30 Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.973549 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tsjd8"] Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.973779 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" podUID="93676336-30fa-466d-8480-8fecf5d5328e" containerName="marketplace-operator" containerID="cri-o://952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62" gracePeriod=30 Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.980300 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpc6s"] Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.980536 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lpc6s" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="registry-server" containerID="cri-o://1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454" gracePeriod=30 Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.991451 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9bblw"] Jan 29 15:07:17 crc kubenswrapper[4767]: I0129 15:07:17.991701 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9bblw" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="registry-server" containerID="cri-o://285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a" gracePeriod=30 Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.009014 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ftjvt"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.009821 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.024125 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ftjvt"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.115213 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx79c\" (UniqueName: \"kubernetes.io/projected/0714bcea-20bf-4faf-b4b1-84df642268ac-kube-api-access-hx79c\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.115269 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.115326 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.216210 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx79c\" (UniqueName: \"kubernetes.io/projected/0714bcea-20bf-4faf-b4b1-84df642268ac-kube-api-access-hx79c\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.216246 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.216282 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.217678 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.222322 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.233887 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx79c\" (UniqueName: \"kubernetes.io/projected/0714bcea-20bf-4faf-b4b1-84df642268ac-kube-api-access-hx79c\") pod \"marketplace-operator-79b997595-ftjvt\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.394301 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.401125 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.462809 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.474689 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.483944 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.523850 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.527379 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-utilities\") pod \"223f99d9-32ef-4685-9eba-b34f09b337b8\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.527504 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztwxr\" (UniqueName: \"kubernetes.io/projected/223f99d9-32ef-4685-9eba-b34f09b337b8-kube-api-access-ztwxr\") pod \"223f99d9-32ef-4685-9eba-b34f09b337b8\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.527581 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-catalog-content\") pod \"223f99d9-32ef-4685-9eba-b34f09b337b8\" (UID: \"223f99d9-32ef-4685-9eba-b34f09b337b8\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.529674 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-utilities" (OuterVolumeSpecName: "utilities") pod "223f99d9-32ef-4685-9eba-b34f09b337b8" (UID: "223f99d9-32ef-4685-9eba-b34f09b337b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.534611 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/223f99d9-32ef-4685-9eba-b34f09b337b8-kube-api-access-ztwxr" (OuterVolumeSpecName: "kube-api-access-ztwxr") pod "223f99d9-32ef-4685-9eba-b34f09b337b8" (UID: "223f99d9-32ef-4685-9eba-b34f09b337b8"). InnerVolumeSpecName "kube-api-access-ztwxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.589831 4767 generic.go:334] "Generic (PLEG): container finished" podID="503997f6-0b05-410b-8011-03f921c6229b" containerID="285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a" exitCode=0 Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.589921 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9bblw" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.589883 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9bblw" event={"ID":"503997f6-0b05-410b-8011-03f921c6229b","Type":"ContainerDied","Data":"285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.590579 4767 scope.go:117] "RemoveContainer" containerID="285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.590810 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9bblw" event={"ID":"503997f6-0b05-410b-8011-03f921c6229b","Type":"ContainerDied","Data":"cd4da699fb7d8cbb0096b9259c1ae39308181ffc93b1c16222ba961f447e9a64"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.594516 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "223f99d9-32ef-4685-9eba-b34f09b337b8" (UID: "223f99d9-32ef-4685-9eba-b34f09b337b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.594564 4767 generic.go:334] "Generic (PLEG): container finished" podID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerID="caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650" exitCode=0 Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.594637 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fjjw" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.594645 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fjjw" event={"ID":"223f99d9-32ef-4685-9eba-b34f09b337b8","Type":"ContainerDied","Data":"caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.594687 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fjjw" event={"ID":"223f99d9-32ef-4685-9eba-b34f09b337b8","Type":"ContainerDied","Data":"212b10bfda4ef226278c6b0bfa7a744ec4e4fefb8b3fc99173720202c7dfa8d5"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.608729 4767 generic.go:334] "Generic (PLEG): container finished" podID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerID="1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454" exitCode=0 Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.608806 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lpc6s" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.608809 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpc6s" event={"ID":"0bed5774-aa4c-4154-9925-6ed6494ca078","Type":"ContainerDied","Data":"1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.608951 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lpc6s" event={"ID":"0bed5774-aa4c-4154-9925-6ed6494ca078","Type":"ContainerDied","Data":"0de971b7dfb5f72e40142ba5d7131dc8c8dd664786582dbf3e3e5b4a8903c6e9"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.611204 4767 generic.go:334] "Generic (PLEG): container finished" podID="77c7c371-0113-48f8-b514-0b53a9671aad" containerID="249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc" exitCode=0 Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.611247 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7srb" event={"ID":"77c7c371-0113-48f8-b514-0b53a9671aad","Type":"ContainerDied","Data":"249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.611266 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s7srb" event={"ID":"77c7c371-0113-48f8-b514-0b53a9671aad","Type":"ContainerDied","Data":"b57c1bb01496f26cd6d6db7020364ea715adea92086e2da0a01519214c9c9401"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.611317 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s7srb" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.613210 4767 scope.go:117] "RemoveContainer" containerID="dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.613839 4767 generic.go:334] "Generic (PLEG): container finished" podID="93676336-30fa-466d-8480-8fecf5d5328e" containerID="952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62" exitCode=0 Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.613872 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" event={"ID":"93676336-30fa-466d-8480-8fecf5d5328e","Type":"ContainerDied","Data":"952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.613948 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" event={"ID":"93676336-30fa-466d-8480-8fecf5d5328e","Type":"ContainerDied","Data":"bfd32486a6dc7b2d394f284ea303509775e4a792b10d97bdaa901e1fbee645ca"} Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.614039 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tsjd8" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629453 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6nn8\" (UniqueName: \"kubernetes.io/projected/93676336-30fa-466d-8480-8fecf5d5328e-kube-api-access-b6nn8\") pod \"93676336-30fa-466d-8480-8fecf5d5328e\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629530 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-catalog-content\") pod \"77c7c371-0113-48f8-b514-0b53a9671aad\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629604 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-catalog-content\") pod \"503997f6-0b05-410b-8011-03f921c6229b\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629636 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-utilities\") pod \"77c7c371-0113-48f8-b514-0b53a9671aad\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629683 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-trusted-ca\") pod \"93676336-30fa-466d-8480-8fecf5d5328e\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629715 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-utilities\") pod \"0bed5774-aa4c-4154-9925-6ed6494ca078\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629762 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-operator-metrics\") pod \"93676336-30fa-466d-8480-8fecf5d5328e\" (UID: \"93676336-30fa-466d-8480-8fecf5d5328e\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629791 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cpnx\" (UniqueName: \"kubernetes.io/projected/0bed5774-aa4c-4154-9925-6ed6494ca078-kube-api-access-9cpnx\") pod \"0bed5774-aa4c-4154-9925-6ed6494ca078\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629827 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-utilities\") pod \"503997f6-0b05-410b-8011-03f921c6229b\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629875 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-catalog-content\") pod \"0bed5774-aa4c-4154-9925-6ed6494ca078\" (UID: \"0bed5774-aa4c-4154-9925-6ed6494ca078\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.629948 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t7b7\" (UniqueName: \"kubernetes.io/projected/503997f6-0b05-410b-8011-03f921c6229b-kube-api-access-5t7b7\") pod \"503997f6-0b05-410b-8011-03f921c6229b\" (UID: \"503997f6-0b05-410b-8011-03f921c6229b\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.630018 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qnth\" (UniqueName: \"kubernetes.io/projected/77c7c371-0113-48f8-b514-0b53a9671aad-kube-api-access-9qnth\") pod \"77c7c371-0113-48f8-b514-0b53a9671aad\" (UID: \"77c7c371-0113-48f8-b514-0b53a9671aad\") " Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.630355 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.630375 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztwxr\" (UniqueName: \"kubernetes.io/projected/223f99d9-32ef-4685-9eba-b34f09b337b8-kube-api-access-ztwxr\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.630392 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/223f99d9-32ef-4685-9eba-b34f09b337b8-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.631861 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-utilities" (OuterVolumeSpecName: "utilities") pod "503997f6-0b05-410b-8011-03f921c6229b" (UID: "503997f6-0b05-410b-8011-03f921c6229b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.633307 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "93676336-30fa-466d-8480-8fecf5d5328e" (UID: "93676336-30fa-466d-8480-8fecf5d5328e"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.635766 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "93676336-30fa-466d-8480-8fecf5d5328e" (UID: "93676336-30fa-466d-8480-8fecf5d5328e"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.637403 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-utilities" (OuterVolumeSpecName: "utilities") pod "77c7c371-0113-48f8-b514-0b53a9671aad" (UID: "77c7c371-0113-48f8-b514-0b53a9671aad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.638940 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-utilities" (OuterVolumeSpecName: "utilities") pod "0bed5774-aa4c-4154-9925-6ed6494ca078" (UID: "0bed5774-aa4c-4154-9925-6ed6494ca078"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.639015 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bed5774-aa4c-4154-9925-6ed6494ca078-kube-api-access-9cpnx" (OuterVolumeSpecName: "kube-api-access-9cpnx") pod "0bed5774-aa4c-4154-9925-6ed6494ca078" (UID: "0bed5774-aa4c-4154-9925-6ed6494ca078"). InnerVolumeSpecName "kube-api-access-9cpnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.646694 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93676336-30fa-466d-8480-8fecf5d5328e-kube-api-access-b6nn8" (OuterVolumeSpecName: "kube-api-access-b6nn8") pod "93676336-30fa-466d-8480-8fecf5d5328e" (UID: "93676336-30fa-466d-8480-8fecf5d5328e"). InnerVolumeSpecName "kube-api-access-b6nn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.647143 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4fjjw"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.650790 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4fjjw"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.652008 4767 scope.go:117] "RemoveContainer" containerID="c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.655328 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/503997f6-0b05-410b-8011-03f921c6229b-kube-api-access-5t7b7" (OuterVolumeSpecName: "kube-api-access-5t7b7") pod "503997f6-0b05-410b-8011-03f921c6229b" (UID: "503997f6-0b05-410b-8011-03f921c6229b"). InnerVolumeSpecName "kube-api-access-5t7b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.657129 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77c7c371-0113-48f8-b514-0b53a9671aad-kube-api-access-9qnth" (OuterVolumeSpecName: "kube-api-access-9qnth") pod "77c7c371-0113-48f8-b514-0b53a9671aad" (UID: "77c7c371-0113-48f8-b514-0b53a9671aad"). InnerVolumeSpecName "kube-api-access-9qnth". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.673356 4767 scope.go:117] "RemoveContainer" containerID="285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.674021 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a\": container with ID starting with 285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a not found: ID does not exist" containerID="285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.674079 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a"} err="failed to get container status \"285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a\": rpc error: code = NotFound desc = could not find container \"285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a\": container with ID starting with 285b94b9bafec18357721f60a4df02a2099807630cb0b9b80c0d4394802d285a not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.674118 4767 scope.go:117] "RemoveContainer" containerID="dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.674865 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe\": container with ID starting with dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe not found: ID does not exist" containerID="dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.674912 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe"} err="failed to get container status \"dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe\": rpc error: code = NotFound desc = could not find container \"dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe\": container with ID starting with dac01b297539449ae400c0ebf75b9e926ed9e976f95824ad7b6da9b7591da3fe not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.674938 4767 scope.go:117] "RemoveContainer" containerID="c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.675394 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50\": container with ID starting with c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50 not found: ID does not exist" containerID="c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.675423 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50"} err="failed to get container status \"c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50\": rpc error: code = NotFound desc = could not find container \"c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50\": container with ID starting with c0477aa440b5f5417cbd4b1fcff12fcf418053e8cf010436b7c5d77923944e50 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.675441 4767 scope.go:117] "RemoveContainer" containerID="caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.681654 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bed5774-aa4c-4154-9925-6ed6494ca078" (UID: "0bed5774-aa4c-4154-9925-6ed6494ca078"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.691760 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77c7c371-0113-48f8-b514-0b53a9671aad" (UID: "77c7c371-0113-48f8-b514-0b53a9671aad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.706069 4767 scope.go:117] "RemoveContainer" containerID="143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731238 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qnth\" (UniqueName: \"kubernetes.io/projected/77c7c371-0113-48f8-b514-0b53a9671aad-kube-api-access-9qnth\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731273 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6nn8\" (UniqueName: \"kubernetes.io/projected/93676336-30fa-466d-8480-8fecf5d5328e-kube-api-access-b6nn8\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731283 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731293 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77c7c371-0113-48f8-b514-0b53a9671aad-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731303 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731311 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731319 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/93676336-30fa-466d-8480-8fecf5d5328e-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731331 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cpnx\" (UniqueName: \"kubernetes.io/projected/0bed5774-aa4c-4154-9925-6ed6494ca078-kube-api-access-9cpnx\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731339 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731346 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed5774-aa4c-4154-9925-6ed6494ca078-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.731355 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t7b7\" (UniqueName: \"kubernetes.io/projected/503997f6-0b05-410b-8011-03f921c6229b-kube-api-access-5t7b7\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.734187 4767 scope.go:117] "RemoveContainer" containerID="750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.748332 4767 scope.go:117] "RemoveContainer" containerID="caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.748807 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650\": container with ID starting with caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650 not found: ID does not exist" containerID="caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.748857 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650"} err="failed to get container status \"caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650\": rpc error: code = NotFound desc = could not find container \"caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650\": container with ID starting with caee89c3a696246f6fa3eeef430fd56fceee26149f050f99fa79da9cd7e77650 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.748886 4767 scope.go:117] "RemoveContainer" containerID="143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.749521 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241\": container with ID starting with 143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241 not found: ID does not exist" containerID="143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.749565 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241"} err="failed to get container status \"143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241\": rpc error: code = NotFound desc = could not find container \"143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241\": container with ID starting with 143b6aa264594a359e96da008de124483621e318d7691078de32b7926b29d241 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.749589 4767 scope.go:117] "RemoveContainer" containerID="750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.749864 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3\": container with ID starting with 750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3 not found: ID does not exist" containerID="750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.749887 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3"} err="failed to get container status \"750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3\": rpc error: code = NotFound desc = could not find container \"750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3\": container with ID starting with 750cf480b5ae11e2295afc68524c3b55671c749c4e8dc91b13f6ed0ed9244df3 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.749917 4767 scope.go:117] "RemoveContainer" containerID="1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.764118 4767 scope.go:117] "RemoveContainer" containerID="7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.779507 4767 scope.go:117] "RemoveContainer" containerID="d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.814633 4767 scope.go:117] "RemoveContainer" containerID="1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.818103 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454\": container with ID starting with 1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454 not found: ID does not exist" containerID="1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.818133 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454"} err="failed to get container status \"1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454\": rpc error: code = NotFound desc = could not find container \"1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454\": container with ID starting with 1ed3be45247b2db31a03606d3972ce5766d90bb7a8cb4e2b75d794001198a454 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.818157 4767 scope.go:117] "RemoveContainer" containerID="7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.819930 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c\": container with ID starting with 7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c not found: ID does not exist" containerID="7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.819953 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c"} err="failed to get container status \"7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c\": rpc error: code = NotFound desc = could not find container \"7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c\": container with ID starting with 7e2dc00cdc2363fa69db0e13819efd3cdcdd6d076487ed99a275c2237a2cb68c not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.819975 4767 scope.go:117] "RemoveContainer" containerID="d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.820155 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c\": container with ID starting with d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c not found: ID does not exist" containerID="d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.820173 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c"} err="failed to get container status \"d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c\": rpc error: code = NotFound desc = could not find container \"d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c\": container with ID starting with d8cf405243792652b4b65a5e43c24d3d6ae0b657e697ec0e866aa49652e7654c not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.820186 4767 scope.go:117] "RemoveContainer" containerID="249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.832304 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j8frh"] Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.832924 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.833003 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.833085 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.833149 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.833214 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.833406 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.833517 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.833629 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.840277 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.840463 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.840529 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.840603 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.840661 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.840712 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.840768 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.840819 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.841521 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93676336-30fa-466d-8480-8fecf5d5328e" containerName="marketplace-operator" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.841625 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="93676336-30fa-466d-8480-8fecf5d5328e" containerName="marketplace-operator" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.841692 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.841750 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="extract-content" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.841808 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.841864 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.841946 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842003 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="extract-utilities" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.842060 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842118 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842313 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842378 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842435 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842487 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="93676336-30fa-466d-8480-8fecf5d5328e" containerName="marketplace-operator" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842550 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="503997f6-0b05-410b-8011-03f921c6229b" containerName="registry-server" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.842960 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j8frh"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.843108 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.844678 4767 scope.go:117] "RemoveContainer" containerID="7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.849496 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "503997f6-0b05-410b-8011-03f921c6229b" (UID: "503997f6-0b05-410b-8011-03f921c6229b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.867715 4767 scope.go:117] "RemoveContainer" containerID="ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.882739 4767 scope.go:117] "RemoveContainer" containerID="249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.883166 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc\": container with ID starting with 249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc not found: ID does not exist" containerID="249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.883215 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc"} err="failed to get container status \"249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc\": rpc error: code = NotFound desc = could not find container \"249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc\": container with ID starting with 249c4465a5695c354b3ec670f3b795da0a49aca68a03b1aa1b9b97e83aa323cc not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.883249 4767 scope.go:117] "RemoveContainer" containerID="7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.883699 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3\": container with ID starting with 7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3 not found: ID does not exist" containerID="7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.883754 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3"} err="failed to get container status \"7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3\": rpc error: code = NotFound desc = could not find container \"7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3\": container with ID starting with 7a13e9a9195d95b9d48f754c4c4d621adaae54dc89141a74128ffe1e5560b0f3 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.883778 4767 scope.go:117] "RemoveContainer" containerID="ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.884140 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9\": container with ID starting with ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9 not found: ID does not exist" containerID="ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.884284 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9"} err="failed to get container status \"ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9\": rpc error: code = NotFound desc = could not find container \"ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9\": container with ID starting with ea3fa5bfab1f6549cfc949d3b2cd7720b2174a4d49d8418ad3a19071287249f9 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.884408 4767 scope.go:117] "RemoveContainer" containerID="952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.897999 4767 scope.go:117] "RemoveContainer" containerID="952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62" Jan 29 15:07:18 crc kubenswrapper[4767]: E0129 15:07:18.898460 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62\": container with ID starting with 952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62 not found: ID does not exist" containerID="952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.898509 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62"} err="failed to get container status \"952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62\": rpc error: code = NotFound desc = could not find container \"952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62\": container with ID starting with 952c84885a3495e280eda6440732ff6b1e2ea52177b4539ba163cfb0a0b37d62 not found: ID does not exist" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.937365 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfvkp\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-kube-api-access-tfvkp\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.937417 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.937453 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-bound-sa-token\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.937481 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.937505 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-registry-certificates\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.938611 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-registry-tls\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.938674 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.938732 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-trusted-ca\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.938834 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/503997f6-0b05-410b-8011-03f921c6229b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.948556 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9bblw"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.959673 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9bblw"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.966557 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ftjvt"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.971690 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tsjd8"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.979031 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tsjd8"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.981879 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.985547 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpc6s"] Jan 29 15:07:18 crc kubenswrapper[4767]: I0129 15:07:18.997329 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lpc6s"] Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.002014 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s7srb"] Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.007867 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s7srb"] Jan 29 15:07:19 crc kubenswrapper[4767]: E0129 15:07:19.011005 4767 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77c7c371_0113_48f8_b514_0b53a9671aad.slice/crio-b57c1bb01496f26cd6d6db7020364ea715adea92086e2da0a01519214c9c9401\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod503997f6_0b05_410b_8011_03f921c6229b.slice/crio-cd4da699fb7d8cbb0096b9259c1ae39308181ffc93b1c16222ba961f447e9a64\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93676336_30fa_466d_8480_8fecf5d5328e.slice/crio-bfd32486a6dc7b2d394f284ea303509775e4a792b10d97bdaa901e1fbee645ca\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77c7c371_0113_48f8_b514_0b53a9671aad.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0bed5774_aa4c_4154_9925_6ed6494ca078.slice\": RecentStats: unable to find data in memory cache]" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.025667 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bed5774-aa4c-4154-9925-6ed6494ca078" path="/var/lib/kubelet/pods/0bed5774-aa4c-4154-9925-6ed6494ca078/volumes" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.026347 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="223f99d9-32ef-4685-9eba-b34f09b337b8" path="/var/lib/kubelet/pods/223f99d9-32ef-4685-9eba-b34f09b337b8/volumes" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.026942 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="503997f6-0b05-410b-8011-03f921c6229b" path="/var/lib/kubelet/pods/503997f6-0b05-410b-8011-03f921c6229b/volumes" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.028055 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77c7c371-0113-48f8-b514-0b53a9671aad" path="/var/lib/kubelet/pods/77c7c371-0113-48f8-b514-0b53a9671aad/volumes" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.029093 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93676336-30fa-466d-8480-8fecf5d5328e" path="/var/lib/kubelet/pods/93676336-30fa-466d-8480-8fecf5d5328e/volumes" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.040047 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-bound-sa-token\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.040100 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-registry-certificates\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.040143 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.040166 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-registry-tls\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.040193 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-trusted-ca\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.040237 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfvkp\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-kube-api-access-tfvkp\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.040263 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.042383 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-trusted-ca\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.043573 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-registry-certificates\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.043769 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-ca-trust-extracted\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.044443 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-installation-pull-secrets\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.045079 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-registry-tls\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.070884 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfvkp\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-kube-api-access-tfvkp\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.071230 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42cf7ebf-7a06-4c3e-bbc6-7776b3259361-bound-sa-token\") pod \"image-registry-66df7c8f76-j8frh\" (UID: \"42cf7ebf-7a06-4c3e-bbc6-7776b3259361\") " pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.163144 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.535291 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-j8frh"] Jan 29 15:07:19 crc kubenswrapper[4767]: W0129 15:07:19.542007 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42cf7ebf_7a06_4c3e_bbc6_7776b3259361.slice/crio-99ae535c6a31dc7eb04632491b85e616c91274bfe8a12e9adfeccbfbc83f6865 WatchSource:0}: Error finding container 99ae535c6a31dc7eb04632491b85e616c91274bfe8a12e9adfeccbfbc83f6865: Status 404 returned error can't find the container with id 99ae535c6a31dc7eb04632491b85e616c91274bfe8a12e9adfeccbfbc83f6865 Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.620620 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" event={"ID":"42cf7ebf-7a06-4c3e-bbc6-7776b3259361","Type":"ContainerStarted","Data":"99ae535c6a31dc7eb04632491b85e616c91274bfe8a12e9adfeccbfbc83f6865"} Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.621833 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" event={"ID":"0714bcea-20bf-4faf-b4b1-84df642268ac","Type":"ContainerStarted","Data":"0b0b514a641d7663293829194c411d5f37d5f592b70affcb95faaeb03829e274"} Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.621931 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" event={"ID":"0714bcea-20bf-4faf-b4b1-84df642268ac","Type":"ContainerStarted","Data":"f53fa6aa6b82d2959119f220c20968a71054c7f12e182f6eb0352bf19df451ae"} Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.623486 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.630189 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:07:19 crc kubenswrapper[4767]: I0129 15:07:19.639536 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" podStartSLOduration=2.63951742 podStartE2EDuration="2.63951742s" podCreationTimestamp="2026-01-29 15:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:07:19.636764909 +0000 UTC m=+395.447081948" watchObservedRunningTime="2026-01-29 15:07:19.63951742 +0000 UTC m=+395.449834459" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.158251 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2975f"] Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.161047 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.167609 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.171111 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2975f"] Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.256249 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-utilities\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.256803 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v25cq\" (UniqueName: \"kubernetes.io/projected/269bdfc1-3775-468a-a982-eebda8f7346e-kube-api-access-v25cq\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.256978 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-catalog-content\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.358428 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-utilities\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.358493 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v25cq\" (UniqueName: \"kubernetes.io/projected/269bdfc1-3775-468a-a982-eebda8f7346e-kube-api-access-v25cq\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.358528 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-catalog-content\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.359004 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-utilities\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.359029 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-catalog-content\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.359863 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-glb2r"] Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.361195 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.363458 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.377486 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glb2r"] Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.381400 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v25cq\" (UniqueName: \"kubernetes.io/projected/269bdfc1-3775-468a-a982-eebda8f7346e-kube-api-access-v25cq\") pod \"redhat-marketplace-2975f\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.459541 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-catalog-content\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.459604 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdxfw\" (UniqueName: \"kubernetes.io/projected/0039b26d-8fc9-483e-9655-8ce3d9048727-kube-api-access-tdxfw\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.459652 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-utilities\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.485010 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.560805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-catalog-content\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.560888 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdxfw\" (UniqueName: \"kubernetes.io/projected/0039b26d-8fc9-483e-9655-8ce3d9048727-kube-api-access-tdxfw\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.560967 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-utilities\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.561384 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-catalog-content\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.561422 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-utilities\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.592529 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdxfw\" (UniqueName: \"kubernetes.io/projected/0039b26d-8fc9-483e-9655-8ce3d9048727-kube-api-access-tdxfw\") pod \"certified-operators-glb2r\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.632757 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" event={"ID":"42cf7ebf-7a06-4c3e-bbc6-7776b3259361","Type":"ContainerStarted","Data":"9125d208145ae3d80b3da75ba3d8b96f62646ea0ac01dfa031553960acbd2eff"} Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.657262 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" podStartSLOduration=2.657241962 podStartE2EDuration="2.657241962s" podCreationTimestamp="2026-01-29 15:07:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:07:20.655410368 +0000 UTC m=+396.465727407" watchObservedRunningTime="2026-01-29 15:07:20.657241962 +0000 UTC m=+396.467559001" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.676957 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:07:20 crc kubenswrapper[4767]: I0129 15:07:20.956458 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2975f"] Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.159163 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-glb2r"] Jan 29 15:07:21 crc kubenswrapper[4767]: W0129 15:07:21.164917 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0039b26d_8fc9_483e_9655_8ce3d9048727.slice/crio-e8bf714f123467db3df8c5397daf88077fd8bcfad5fd80cc02f35229aa4f5f78 WatchSource:0}: Error finding container e8bf714f123467db3df8c5397daf88077fd8bcfad5fd80cc02f35229aa4f5f78: Status 404 returned error can't find the container with id e8bf714f123467db3df8c5397daf88077fd8bcfad5fd80cc02f35229aa4f5f78 Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.640207 4767 generic.go:334] "Generic (PLEG): container finished" podID="269bdfc1-3775-468a-a982-eebda8f7346e" containerID="2a37e58d2297c3d114f228f518857857c919fc3b13486f89629b238aae8d3f3f" exitCode=0 Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.640298 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2975f" event={"ID":"269bdfc1-3775-468a-a982-eebda8f7346e","Type":"ContainerDied","Data":"2a37e58d2297c3d114f228f518857857c919fc3b13486f89629b238aae8d3f3f"} Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.640332 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2975f" event={"ID":"269bdfc1-3775-468a-a982-eebda8f7346e","Type":"ContainerStarted","Data":"bd8f0028e3046fdb498c3509a9644df610648749dd5640929da6fec8b36cd823"} Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.645155 4767 generic.go:334] "Generic (PLEG): container finished" podID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerID="3498a85f1679f1407544bfa63dfc8cffde177a7f016e3cee801d02b5a0e15cd1" exitCode=0 Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.645299 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glb2r" event={"ID":"0039b26d-8fc9-483e-9655-8ce3d9048727","Type":"ContainerDied","Data":"3498a85f1679f1407544bfa63dfc8cffde177a7f016e3cee801d02b5a0e15cd1"} Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.645383 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glb2r" event={"ID":"0039b26d-8fc9-483e-9655-8ce3d9048727","Type":"ContainerStarted","Data":"e8bf714f123467db3df8c5397daf88077fd8bcfad5fd80cc02f35229aa4f5f78"} Jan 29 15:07:21 crc kubenswrapper[4767]: I0129 15:07:21.645594 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:21 crc kubenswrapper[4767]: E0129 15:07:21.769049 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:07:21 crc kubenswrapper[4767]: E0129 15:07:21.769483 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tdxfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-glb2r_openshift-marketplace(0039b26d-8fc9-483e-9655-8ce3d9048727): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:21 crc kubenswrapper[4767]: E0129 15:07:21.770888 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" Jan 29 15:07:21 crc kubenswrapper[4767]: E0129 15:07:21.787706 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:07:21 crc kubenswrapper[4767]: E0129 15:07:21.787836 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v25cq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2975f_openshift-marketplace(269bdfc1-3775-468a-a982-eebda8f7346e): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:21 crc kubenswrapper[4767]: E0129 15:07:21.789005 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" Jan 29 15:07:22 crc kubenswrapper[4767]: E0129 15:07:22.652662 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" Jan 29 15:07:22 crc kubenswrapper[4767]: E0129 15:07:22.654254 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.760126 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-24ntq"] Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.761574 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.763451 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.765056 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-24ntq"] Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.907404 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62mwb\" (UniqueName: \"kubernetes.io/projected/8579daf0-a4b5-473c-9133-048eaa57c5bb-kube-api-access-62mwb\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.907490 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-utilities\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.907524 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-catalog-content\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.960684 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tp7wv"] Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.961847 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.963714 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 15:07:22 crc kubenswrapper[4767]: I0129 15:07:22.969163 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp7wv"] Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.008827 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62mwb\" (UniqueName: \"kubernetes.io/projected/8579daf0-a4b5-473c-9133-048eaa57c5bb-kube-api-access-62mwb\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.008916 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-utilities\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.008954 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-catalog-content\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.009534 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-catalog-content\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.009646 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-utilities\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.027002 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62mwb\" (UniqueName: \"kubernetes.io/projected/8579daf0-a4b5-473c-9133-048eaa57c5bb-kube-api-access-62mwb\") pod \"redhat-operators-24ntq\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.110254 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-utilities\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.110385 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-catalog-content\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.110434 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffkxn\" (UniqueName: \"kubernetes.io/projected/87659dc2-6bea-4599-8131-eaa87aee2406-kube-api-access-ffkxn\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.133547 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.211734 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-catalog-content\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.211774 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffkxn\" (UniqueName: \"kubernetes.io/projected/87659dc2-6bea-4599-8131-eaa87aee2406-kube-api-access-ffkxn\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.211841 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-utilities\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.212605 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-catalog-content\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.213425 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-utilities\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.230446 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffkxn\" (UniqueName: \"kubernetes.io/projected/87659dc2-6bea-4599-8131-eaa87aee2406-kube-api-access-ffkxn\") pod \"community-operators-tp7wv\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.282072 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.511977 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-24ntq"] Jan 29 15:07:23 crc kubenswrapper[4767]: W0129 15:07:23.516777 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8579daf0_a4b5_473c_9133_048eaa57c5bb.slice/crio-d9663fe41250c3c6c980ea39b239067dcb1596918b5d6f1517b555675ce30a3d WatchSource:0}: Error finding container d9663fe41250c3c6c980ea39b239067dcb1596918b5d6f1517b555675ce30a3d: Status 404 returned error can't find the container with id d9663fe41250c3c6c980ea39b239067dcb1596918b5d6f1517b555675ce30a3d Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.661791 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerStarted","Data":"1cc1111c5e3e0566a4412462e20991b0e247b95575f38cb22ba51af108a83ce5"} Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.661852 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerStarted","Data":"d9663fe41250c3c6c980ea39b239067dcb1596918b5d6f1517b555675ce30a3d"} Jan 29 15:07:23 crc kubenswrapper[4767]: I0129 15:07:23.682539 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tp7wv"] Jan 29 15:07:23 crc kubenswrapper[4767]: W0129 15:07:23.717512 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87659dc2_6bea_4599_8131_eaa87aee2406.slice/crio-1bf083074d437ac09c0afe62491036cc5911996756aaa65b39db68f76fd58150 WatchSource:0}: Error finding container 1bf083074d437ac09c0afe62491036cc5911996756aaa65b39db68f76fd58150: Status 404 returned error can't find the container with id 1bf083074d437ac09c0afe62491036cc5911996756aaa65b39db68f76fd58150 Jan 29 15:07:23 crc kubenswrapper[4767]: E0129 15:07:23.796698 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:07:23 crc kubenswrapper[4767]: E0129 15:07:23.797100 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-62mwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-24ntq_openshift-marketplace(8579daf0-a4b5-473c-9133-048eaa57c5bb): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:23 crc kubenswrapper[4767]: E0129 15:07:23.798317 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:07:24 crc kubenswrapper[4767]: I0129 15:07:24.668496 4767 generic.go:334] "Generic (PLEG): container finished" podID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerID="1cc1111c5e3e0566a4412462e20991b0e247b95575f38cb22ba51af108a83ce5" exitCode=0 Jan 29 15:07:24 crc kubenswrapper[4767]: I0129 15:07:24.668598 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerDied","Data":"1cc1111c5e3e0566a4412462e20991b0e247b95575f38cb22ba51af108a83ce5"} Jan 29 15:07:24 crc kubenswrapper[4767]: I0129 15:07:24.670867 4767 generic.go:334] "Generic (PLEG): container finished" podID="87659dc2-6bea-4599-8131-eaa87aee2406" containerID="d0fdf6efa5c6c883084f5ba572b581986b85c832f766c65f70b40715dabaa58c" exitCode=0 Jan 29 15:07:24 crc kubenswrapper[4767]: I0129 15:07:24.671015 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp7wv" event={"ID":"87659dc2-6bea-4599-8131-eaa87aee2406","Type":"ContainerDied","Data":"d0fdf6efa5c6c883084f5ba572b581986b85c832f766c65f70b40715dabaa58c"} Jan 29 15:07:24 crc kubenswrapper[4767]: I0129 15:07:24.671067 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp7wv" event={"ID":"87659dc2-6bea-4599-8131-eaa87aee2406","Type":"ContainerStarted","Data":"1bf083074d437ac09c0afe62491036cc5911996756aaa65b39db68f76fd58150"} Jan 29 15:07:24 crc kubenswrapper[4767]: E0129 15:07:24.671738 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:07:24 crc kubenswrapper[4767]: E0129 15:07:24.820171 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:07:24 crc kubenswrapper[4767]: E0129 15:07:24.820327 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffkxn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tp7wv_openshift-marketplace(87659dc2-6bea-4599-8131-eaa87aee2406): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:24 crc kubenswrapper[4767]: E0129 15:07:24.822457 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:07:25 crc kubenswrapper[4767]: E0129 15:07:25.679171 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:07:25 crc kubenswrapper[4767]: E0129 15:07:25.679487 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:07:36 crc kubenswrapper[4767]: E0129 15:07:36.146546 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:07:36 crc kubenswrapper[4767]: E0129 15:07:36.147518 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v25cq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2975f_openshift-marketplace(269bdfc1-3775-468a-a982-eebda8f7346e): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:36 crc kubenswrapper[4767]: E0129 15:07:36.148809 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" Jan 29 15:07:37 crc kubenswrapper[4767]: E0129 15:07:37.158133 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:07:37 crc kubenswrapper[4767]: E0129 15:07:37.158290 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tdxfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-glb2r_openshift-marketplace(0039b26d-8fc9-483e-9655-8ce3d9048727): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:37 crc kubenswrapper[4767]: E0129 15:07:37.159433 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" Jan 29 15:07:39 crc kubenswrapper[4767]: I0129 15:07:39.170087 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-j8frh" Jan 29 15:07:39 crc kubenswrapper[4767]: E0129 15:07:39.239434 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:07:39 crc kubenswrapper[4767]: E0129 15:07:39.239620 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffkxn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tp7wv_openshift-marketplace(87659dc2-6bea-4599-8131-eaa87aee2406): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:39 crc kubenswrapper[4767]: E0129 15:07:39.240726 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:07:39 crc kubenswrapper[4767]: I0129 15:07:39.246069 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rmw5d"] Jan 29 15:07:39 crc kubenswrapper[4767]: I0129 15:07:39.932352 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-5w242"] Jan 29 15:07:39 crc kubenswrapper[4767]: I0129 15:07:39.932616 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" podUID="f9799634-f53e-406a-9918-135ba5f9d891" containerName="controller-manager" containerID="cri-o://aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0" gracePeriod=30 Jan 29 15:07:40 crc kubenswrapper[4767]: E0129 15:07:40.192435 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:07:40 crc kubenswrapper[4767]: E0129 15:07:40.192821 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-62mwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-24ntq_openshift-marketplace(8579daf0-a4b5-473c-9133-048eaa57c5bb): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:07:40 crc kubenswrapper[4767]: E0129 15:07:40.194302 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.322091 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.397279 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gthz\" (UniqueName: \"kubernetes.io/projected/f9799634-f53e-406a-9918-135ba5f9d891-kube-api-access-7gthz\") pod \"f9799634-f53e-406a-9918-135ba5f9d891\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.397339 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-proxy-ca-bundles\") pod \"f9799634-f53e-406a-9918-135ba5f9d891\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.397358 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-config\") pod \"f9799634-f53e-406a-9918-135ba5f9d891\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.397376 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9799634-f53e-406a-9918-135ba5f9d891-serving-cert\") pod \"f9799634-f53e-406a-9918-135ba5f9d891\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.397413 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-client-ca\") pod \"f9799634-f53e-406a-9918-135ba5f9d891\" (UID: \"f9799634-f53e-406a-9918-135ba5f9d891\") " Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.398145 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f9799634-f53e-406a-9918-135ba5f9d891" (UID: "f9799634-f53e-406a-9918-135ba5f9d891"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.398204 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-config" (OuterVolumeSpecName: "config") pod "f9799634-f53e-406a-9918-135ba5f9d891" (UID: "f9799634-f53e-406a-9918-135ba5f9d891"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.398217 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-client-ca" (OuterVolumeSpecName: "client-ca") pod "f9799634-f53e-406a-9918-135ba5f9d891" (UID: "f9799634-f53e-406a-9918-135ba5f9d891"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.403615 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9799634-f53e-406a-9918-135ba5f9d891-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f9799634-f53e-406a-9918-135ba5f9d891" (UID: "f9799634-f53e-406a-9918-135ba5f9d891"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.404355 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9799634-f53e-406a-9918-135ba5f9d891-kube-api-access-7gthz" (OuterVolumeSpecName: "kube-api-access-7gthz") pod "f9799634-f53e-406a-9918-135ba5f9d891" (UID: "f9799634-f53e-406a-9918-135ba5f9d891"). InnerVolumeSpecName "kube-api-access-7gthz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.497919 4767 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.497957 4767 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.497966 4767 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9799634-f53e-406a-9918-135ba5f9d891-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.497975 4767 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9799634-f53e-406a-9918-135ba5f9d891-client-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.497986 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gthz\" (UniqueName: \"kubernetes.io/projected/f9799634-f53e-406a-9918-135ba5f9d891-kube-api-access-7gthz\") on node \"crc\" DevicePath \"\"" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.775341 4767 generic.go:334] "Generic (PLEG): container finished" podID="f9799634-f53e-406a-9918-135ba5f9d891" containerID="aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0" exitCode=0 Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.775407 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" event={"ID":"f9799634-f53e-406a-9918-135ba5f9d891","Type":"ContainerDied","Data":"aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0"} Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.775497 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" event={"ID":"f9799634-f53e-406a-9918-135ba5f9d891","Type":"ContainerDied","Data":"e942b8aee02409782932b0ca3336837d7b4b3d446648892af3437b406e257e85"} Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.775556 4767 scope.go:117] "RemoveContainer" containerID="aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.776562 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658fd5994d-5w242" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.801964 4767 scope.go:117] "RemoveContainer" containerID="aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0" Jan 29 15:07:40 crc kubenswrapper[4767]: E0129 15:07:40.802609 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0\": container with ID starting with aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0 not found: ID does not exist" containerID="aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.802679 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0"} err="failed to get container status \"aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0\": rpc error: code = NotFound desc = could not find container \"aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0\": container with ID starting with aba5f2eae455a0d200c441dfa5dfda36742b8697ddbb4bd291604108a8fcdbe0 not found: ID does not exist" Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.826013 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-5w242"] Jan 29 15:07:40 crc kubenswrapper[4767]: I0129 15:07:40.833300 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-658fd5994d-5w242"] Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.026443 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9799634-f53e-406a-9918-135ba5f9d891" path="/var/lib/kubelet/pods/f9799634-f53e-406a-9918-135ba5f9d891/volumes" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.537819 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-ng4bb"] Jan 29 15:07:41 crc kubenswrapper[4767]: E0129 15:07:41.538047 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9799634-f53e-406a-9918-135ba5f9d891" containerName="controller-manager" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.538064 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9799634-f53e-406a-9918-135ba5f9d891" containerName="controller-manager" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.538176 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9799634-f53e-406a-9918-135ba5f9d891" containerName="controller-manager" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.538654 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.541063 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.541305 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.542383 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.543098 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.543547 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.545377 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.552757 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-ng4bb"] Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.554123 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.714342 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-client-ca\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.715057 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-config\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.715242 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pcqd\" (UniqueName: \"kubernetes.io/projected/b5e054e7-e150-4190-910b-5f95da1c9155-kube-api-access-4pcqd\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.715400 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e054e7-e150-4190-910b-5f95da1c9155-serving-cert\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.715581 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-proxy-ca-bundles\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.817373 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-client-ca\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.818036 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-config\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.818093 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pcqd\" (UniqueName: \"kubernetes.io/projected/b5e054e7-e150-4190-910b-5f95da1c9155-kube-api-access-4pcqd\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.818119 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e054e7-e150-4190-910b-5f95da1c9155-serving-cert\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.818171 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-proxy-ca-bundles\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.818348 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-client-ca\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.819539 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-proxy-ca-bundles\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.819695 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5e054e7-e150-4190-910b-5f95da1c9155-config\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.825292 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5e054e7-e150-4190-910b-5f95da1c9155-serving-cert\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.836151 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pcqd\" (UniqueName: \"kubernetes.io/projected/b5e054e7-e150-4190-910b-5f95da1c9155-kube-api-access-4pcqd\") pod \"controller-manager-66b4b547f8-ng4bb\" (UID: \"b5e054e7-e150-4190-910b-5f95da1c9155\") " pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:41 crc kubenswrapper[4767]: I0129 15:07:41.854404 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.287110 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66b4b547f8-ng4bb"] Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.789723 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" event={"ID":"b5e054e7-e150-4190-910b-5f95da1c9155","Type":"ContainerStarted","Data":"903cd59f85f3fed86ba6bf6280d8210ff620681fecc325da49710a38300e0b1e"} Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.790023 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.790035 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" event={"ID":"b5e054e7-e150-4190-910b-5f95da1c9155","Type":"ContainerStarted","Data":"1d8beb6b7c35ddfcb8951478082ff8ad1f1ec1bde9060c05cc71d28724497601"} Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.794048 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.805080 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.805137 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.805188 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.805800 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2b48f9b4d2a600ad7259d8019bac620bc43f521250d0bb6cd00c64ef387aa854"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.805859 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://2b48f9b4d2a600ad7259d8019bac620bc43f521250d0bb6cd00c64ef387aa854" gracePeriod=600 Jan 29 15:07:42 crc kubenswrapper[4767]: I0129 15:07:42.810231 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" podStartSLOduration=3.810209982 podStartE2EDuration="3.810209982s" podCreationTimestamp="2026-01-29 15:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:07:42.804425822 +0000 UTC m=+418.614742861" watchObservedRunningTime="2026-01-29 15:07:42.810209982 +0000 UTC m=+418.620527021" Jan 29 15:07:43 crc kubenswrapper[4767]: I0129 15:07:43.800993 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="2b48f9b4d2a600ad7259d8019bac620bc43f521250d0bb6cd00c64ef387aa854" exitCode=0 Jan 29 15:07:43 crc kubenswrapper[4767]: I0129 15:07:43.801102 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"2b48f9b4d2a600ad7259d8019bac620bc43f521250d0bb6cd00c64ef387aa854"} Jan 29 15:07:43 crc kubenswrapper[4767]: I0129 15:07:43.803029 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"e4e80bf48c84b6e95f0388dfea83fecc89fce6984a1a924645de1ebde40c9e4c"} Jan 29 15:07:43 crc kubenswrapper[4767]: I0129 15:07:43.803111 4767 scope.go:117] "RemoveContainer" containerID="a32d739a814ac1bd7ff7a6f7fd93555aad6f81897a206d1043d987abf48ff350" Jan 29 15:07:47 crc kubenswrapper[4767]: E0129 15:07:47.032724 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" Jan 29 15:07:48 crc kubenswrapper[4767]: E0129 15:07:48.020947 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" Jan 29 15:07:51 crc kubenswrapper[4767]: E0129 15:07:51.021143 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:07:52 crc kubenswrapper[4767]: E0129 15:07:52.022831 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:08:00 crc kubenswrapper[4767]: E0129 15:08:00.166406 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:08:00 crc kubenswrapper[4767]: E0129 15:08:00.166831 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v25cq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2975f_openshift-marketplace(269bdfc1-3775-468a-a982-eebda8f7346e): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:08:00 crc kubenswrapper[4767]: E0129 15:08:00.168763 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" Jan 29 15:08:03 crc kubenswrapper[4767]: E0129 15:08:03.200944 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:08:03 crc kubenswrapper[4767]: E0129 15:08:03.201457 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tdxfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-glb2r_openshift-marketplace(0039b26d-8fc9-483e-9655-8ce3d9048727): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:08:03 crc kubenswrapper[4767]: E0129 15:08:03.202734 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" Jan 29 15:08:04 crc kubenswrapper[4767]: E0129 15:08:04.145668 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:08:04 crc kubenswrapper[4767]: E0129 15:08:04.145963 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-62mwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-24ntq_openshift-marketplace(8579daf0-a4b5-473c-9133-048eaa57c5bb): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:08:04 crc kubenswrapper[4767]: E0129 15:08:04.147240 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.283709 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" podUID="565418a4-f20c-4841-92fa-d40804d42980" containerName="registry" containerID="cri-o://f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261" gracePeriod=30 Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.789090 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901231 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-bound-sa-token\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901294 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/565418a4-f20c-4841-92fa-d40804d42980-installation-pull-secrets\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901340 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-registry-certificates\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901367 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/565418a4-f20c-4841-92fa-d40804d42980-ca-trust-extracted\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901604 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901669 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-trusted-ca\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901700 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-registry-tls\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.901724 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n6ch\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-kube-api-access-6n6ch\") pod \"565418a4-f20c-4841-92fa-d40804d42980\" (UID: \"565418a4-f20c-4841-92fa-d40804d42980\") " Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.903245 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.903469 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.909190 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565418a4-f20c-4841-92fa-d40804d42980-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.910150 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.912844 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.917107 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.917401 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-kube-api-access-6n6ch" (OuterVolumeSpecName: "kube-api-access-6n6ch") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "kube-api-access-6n6ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.928321 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565418a4-f20c-4841-92fa-d40804d42980-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "565418a4-f20c-4841-92fa-d40804d42980" (UID: "565418a4-f20c-4841-92fa-d40804d42980"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.967255 4767 generic.go:334] "Generic (PLEG): container finished" podID="565418a4-f20c-4841-92fa-d40804d42980" containerID="f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261" exitCode=0 Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.967317 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.967313 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" event={"ID":"565418a4-f20c-4841-92fa-d40804d42980","Type":"ContainerDied","Data":"f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261"} Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.967378 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rmw5d" event={"ID":"565418a4-f20c-4841-92fa-d40804d42980","Type":"ContainerDied","Data":"40f3383acf2557660bac2f696489e1618b9f5a3df1789c04de3377ce685f485d"} Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.967401 4767 scope.go:117] "RemoveContainer" containerID="f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.993510 4767 scope.go:117] "RemoveContainer" containerID="f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261" Jan 29 15:08:04 crc kubenswrapper[4767]: E0129 15:08:04.994836 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261\": container with ID starting with f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261 not found: ID does not exist" containerID="f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.994881 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261"} err="failed to get container status \"f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261\": rpc error: code = NotFound desc = could not find container \"f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261\": container with ID starting with f61b10eadfcf37749795444eb0022cba7eab002674fc5e4be46969363c151261 not found: ID does not exist" Jan 29 15:08:04 crc kubenswrapper[4767]: I0129 15:08:04.998587 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rmw5d"] Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.003628 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.003655 4767 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.003668 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n6ch\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-kube-api-access-6n6ch\") on node \"crc\" DevicePath \"\"" Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.003680 4767 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/565418a4-f20c-4841-92fa-d40804d42980-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.003692 4767 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/565418a4-f20c-4841-92fa-d40804d42980-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.003706 4767 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/565418a4-f20c-4841-92fa-d40804d42980-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.003717 4767 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/565418a4-f20c-4841-92fa-d40804d42980-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.007161 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rmw5d"] Jan 29 15:08:05 crc kubenswrapper[4767]: I0129 15:08:05.026211 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565418a4-f20c-4841-92fa-d40804d42980" path="/var/lib/kubelet/pods/565418a4-f20c-4841-92fa-d40804d42980/volumes" Jan 29 15:08:06 crc kubenswrapper[4767]: E0129 15:08:06.154698 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:08:06 crc kubenswrapper[4767]: E0129 15:08:06.155463 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffkxn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tp7wv_openshift-marketplace(87659dc2-6bea-4599-8131-eaa87aee2406): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:08:06 crc kubenswrapper[4767]: E0129 15:08:06.156660 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:08:13 crc kubenswrapper[4767]: E0129 15:08:13.019766 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" Jan 29 15:08:16 crc kubenswrapper[4767]: E0129 15:08:16.021550 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:08:16 crc kubenswrapper[4767]: E0129 15:08:16.021937 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" Jan 29 15:08:18 crc kubenswrapper[4767]: E0129 15:08:18.020055 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:08:27 crc kubenswrapper[4767]: E0129 15:08:27.020589 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" Jan 29 15:08:28 crc kubenswrapper[4767]: E0129 15:08:28.021257 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:08:29 crc kubenswrapper[4767]: E0129 15:08:29.020723 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" Jan 29 15:08:32 crc kubenswrapper[4767]: E0129 15:08:32.019713 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:08:41 crc kubenswrapper[4767]: E0129 15:08:41.020648 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" Jan 29 15:08:41 crc kubenswrapper[4767]: I0129 15:08:41.021582 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:08:42 crc kubenswrapper[4767]: I0129 15:08:42.204814 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2975f" event={"ID":"269bdfc1-3775-468a-a982-eebda8f7346e","Type":"ContainerStarted","Data":"8401f3dc733c542521518b734aaaf0b514baf17d2d9948b954ca6c68e13f7431"} Jan 29 15:08:43 crc kubenswrapper[4767]: I0129 15:08:43.213157 4767 generic.go:334] "Generic (PLEG): container finished" podID="269bdfc1-3775-468a-a982-eebda8f7346e" containerID="8401f3dc733c542521518b734aaaf0b514baf17d2d9948b954ca6c68e13f7431" exitCode=0 Jan 29 15:08:43 crc kubenswrapper[4767]: I0129 15:08:43.213276 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2975f" event={"ID":"269bdfc1-3775-468a-a982-eebda8f7346e","Type":"ContainerDied","Data":"8401f3dc733c542521518b734aaaf0b514baf17d2d9948b954ca6c68e13f7431"} Jan 29 15:08:45 crc kubenswrapper[4767]: I0129 15:08:45.227598 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2975f" event={"ID":"269bdfc1-3775-468a-a982-eebda8f7346e","Type":"ContainerStarted","Data":"f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d"} Jan 29 15:08:45 crc kubenswrapper[4767]: I0129 15:08:45.246297 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2975f" podStartSLOduration=2.634929681 podStartE2EDuration="1m25.246267455s" podCreationTimestamp="2026-01-29 15:07:20 +0000 UTC" firstStartedPulling="2026-01-29 15:07:21.642949702 +0000 UTC m=+397.453266741" lastFinishedPulling="2026-01-29 15:08:44.254287476 +0000 UTC m=+480.064604515" observedRunningTime="2026-01-29 15:08:45.243071181 +0000 UTC m=+481.053388220" watchObservedRunningTime="2026-01-29 15:08:45.246267455 +0000 UTC m=+481.056584494" Jan 29 15:08:46 crc kubenswrapper[4767]: E0129 15:08:46.021825 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" Jan 29 15:08:46 crc kubenswrapper[4767]: I0129 15:08:46.236415 4767 generic.go:334] "Generic (PLEG): container finished" podID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerID="0bb6129f1feb35a6265aa8cd32b154087d97aaaa46d047f1c65276bcd5be0271" exitCode=0 Jan 29 15:08:46 crc kubenswrapper[4767]: I0129 15:08:46.236487 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glb2r" event={"ID":"0039b26d-8fc9-483e-9655-8ce3d9048727","Type":"ContainerDied","Data":"0bb6129f1feb35a6265aa8cd32b154087d97aaaa46d047f1c65276bcd5be0271"} Jan 29 15:08:48 crc kubenswrapper[4767]: I0129 15:08:48.253515 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glb2r" event={"ID":"0039b26d-8fc9-483e-9655-8ce3d9048727","Type":"ContainerStarted","Data":"e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46"} Jan 29 15:08:48 crc kubenswrapper[4767]: I0129 15:08:48.275675 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-glb2r" podStartSLOduration=2.735364641 podStartE2EDuration="1m28.2756481s" podCreationTimestamp="2026-01-29 15:07:20 +0000 UTC" firstStartedPulling="2026-01-29 15:07:21.64729219 +0000 UTC m=+397.457609229" lastFinishedPulling="2026-01-29 15:08:47.187575649 +0000 UTC m=+482.997892688" observedRunningTime="2026-01-29 15:08:48.269836309 +0000 UTC m=+484.080153348" watchObservedRunningTime="2026-01-29 15:08:48.2756481 +0000 UTC m=+484.085965159" Jan 29 15:08:50 crc kubenswrapper[4767]: I0129 15:08:50.486014 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:08:50 crc kubenswrapper[4767]: I0129 15:08:50.486454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:08:50 crc kubenswrapper[4767]: I0129 15:08:50.535158 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:08:50 crc kubenswrapper[4767]: I0129 15:08:50.678028 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:08:50 crc kubenswrapper[4767]: I0129 15:08:50.678746 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:08:50 crc kubenswrapper[4767]: I0129 15:08:50.722074 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:08:51 crc kubenswrapper[4767]: I0129 15:08:51.319544 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:08:52 crc kubenswrapper[4767]: I0129 15:08:52.321099 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:08:54 crc kubenswrapper[4767]: I0129 15:08:54.291706 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerStarted","Data":"837e6fa44a173a70075a6b7c45bbaa0f698e024f965a5f2bf5d245b37ee6851d"} Jan 29 15:08:55 crc kubenswrapper[4767]: I0129 15:08:55.298799 4767 generic.go:334] "Generic (PLEG): container finished" podID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerID="837e6fa44a173a70075a6b7c45bbaa0f698e024f965a5f2bf5d245b37ee6851d" exitCode=0 Jan 29 15:08:55 crc kubenswrapper[4767]: I0129 15:08:55.298862 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerDied","Data":"837e6fa44a173a70075a6b7c45bbaa0f698e024f965a5f2bf5d245b37ee6851d"} Jan 29 15:08:56 crc kubenswrapper[4767]: I0129 15:08:56.309035 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerStarted","Data":"5425a81c114abd07b63c498185cea3583adbde68d33a2a7d0cafb65b8459b0ae"} Jan 29 15:08:56 crc kubenswrapper[4767]: I0129 15:08:56.331465 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-24ntq" podStartSLOduration=2.144616527 podStartE2EDuration="1m34.33144344s" podCreationTimestamp="2026-01-29 15:07:22 +0000 UTC" firstStartedPulling="2026-01-29 15:07:23.664305859 +0000 UTC m=+399.474622898" lastFinishedPulling="2026-01-29 15:08:55.851132762 +0000 UTC m=+491.661449811" observedRunningTime="2026-01-29 15:08:56.326155504 +0000 UTC m=+492.136472553" watchObservedRunningTime="2026-01-29 15:08:56.33144344 +0000 UTC m=+492.141760479" Jan 29 15:09:00 crc kubenswrapper[4767]: I0129 15:09:00.331178 4767 generic.go:334] "Generic (PLEG): container finished" podID="87659dc2-6bea-4599-8131-eaa87aee2406" containerID="53d1cb912b439b32e9b89ff254f6d2033ccf5272f466035d890cb1c0738c3f4b" exitCode=0 Jan 29 15:09:00 crc kubenswrapper[4767]: I0129 15:09:00.331218 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp7wv" event={"ID":"87659dc2-6bea-4599-8131-eaa87aee2406","Type":"ContainerDied","Data":"53d1cb912b439b32e9b89ff254f6d2033ccf5272f466035d890cb1c0738c3f4b"} Jan 29 15:09:02 crc kubenswrapper[4767]: I0129 15:09:02.347197 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp7wv" event={"ID":"87659dc2-6bea-4599-8131-eaa87aee2406","Type":"ContainerStarted","Data":"2cb2ec7b012335fd18b42138e7d28da61f893b4ec981f29e4acb5ed4d36dfa42"} Jan 29 15:09:02 crc kubenswrapper[4767]: I0129 15:09:02.368548 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tp7wv" podStartSLOduration=3.5384129460000002 podStartE2EDuration="1m40.368528885s" podCreationTimestamp="2026-01-29 15:07:22 +0000 UTC" firstStartedPulling="2026-01-29 15:07:24.672049297 +0000 UTC m=+400.482366336" lastFinishedPulling="2026-01-29 15:09:01.502165236 +0000 UTC m=+497.312482275" observedRunningTime="2026-01-29 15:09:02.366358931 +0000 UTC m=+498.176675970" watchObservedRunningTime="2026-01-29 15:09:02.368528885 +0000 UTC m=+498.178845914" Jan 29 15:09:03 crc kubenswrapper[4767]: I0129 15:09:03.134118 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:09:03 crc kubenswrapper[4767]: I0129 15:09:03.134729 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:09:03 crc kubenswrapper[4767]: I0129 15:09:03.211147 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:09:03 crc kubenswrapper[4767]: I0129 15:09:03.283111 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:09:03 crc kubenswrapper[4767]: I0129 15:09:03.283163 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:09:03 crc kubenswrapper[4767]: I0129 15:09:03.389427 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:09:04 crc kubenswrapper[4767]: I0129 15:09:04.337869 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="registry-server" probeResult="failure" output=< Jan 29 15:09:04 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:09:04 crc kubenswrapper[4767]: > Jan 29 15:09:13 crc kubenswrapper[4767]: I0129 15:09:13.333872 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:09:13 crc kubenswrapper[4767]: I0129 15:09:13.403108 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:10:12 crc kubenswrapper[4767]: I0129 15:10:12.806349 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:10:12 crc kubenswrapper[4767]: I0129 15:10:12.807010 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:10:42 crc kubenswrapper[4767]: I0129 15:10:42.806263 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:10:42 crc kubenswrapper[4767]: I0129 15:10:42.806823 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:11:12 crc kubenswrapper[4767]: I0129 15:11:12.805674 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:11:12 crc kubenswrapper[4767]: I0129 15:11:12.806318 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:11:12 crc kubenswrapper[4767]: I0129 15:11:12.806383 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:11:12 crc kubenswrapper[4767]: I0129 15:11:12.808190 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4e80bf48c84b6e95f0388dfea83fecc89fce6984a1a924645de1ebde40c9e4c"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:11:12 crc kubenswrapper[4767]: I0129 15:11:12.808315 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://e4e80bf48c84b6e95f0388dfea83fecc89fce6984a1a924645de1ebde40c9e4c" gracePeriod=600 Jan 29 15:11:13 crc kubenswrapper[4767]: I0129 15:11:13.088587 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="e4e80bf48c84b6e95f0388dfea83fecc89fce6984a1a924645de1ebde40c9e4c" exitCode=0 Jan 29 15:11:13 crc kubenswrapper[4767]: I0129 15:11:13.088686 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"e4e80bf48c84b6e95f0388dfea83fecc89fce6984a1a924645de1ebde40c9e4c"} Jan 29 15:11:13 crc kubenswrapper[4767]: I0129 15:11:13.089141 4767 scope.go:117] "RemoveContainer" containerID="2b48f9b4d2a600ad7259d8019bac620bc43f521250d0bb6cd00c64ef387aa854" Jan 29 15:11:14 crc kubenswrapper[4767]: I0129 15:11:14.098142 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"dea4eb1d9800844b85bbca49f74990272d70b69b16fee301084f0ca32a79cb1a"} Jan 29 15:13:30 crc kubenswrapper[4767]: I0129 15:13:30.231283 4767 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 29 15:13:42 crc kubenswrapper[4767]: I0129 15:13:42.805642 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:13:42 crc kubenswrapper[4767]: I0129 15:13:42.806273 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:14:12 crc kubenswrapper[4767]: I0129 15:14:12.806255 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:14:12 crc kubenswrapper[4767]: I0129 15:14:12.806997 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:14:42 crc kubenswrapper[4767]: I0129 15:14:42.805844 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:14:42 crc kubenswrapper[4767]: I0129 15:14:42.806486 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:14:42 crc kubenswrapper[4767]: I0129 15:14:42.806542 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:14:42 crc kubenswrapper[4767]: I0129 15:14:42.807300 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dea4eb1d9800844b85bbca49f74990272d70b69b16fee301084f0ca32a79cb1a"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:14:42 crc kubenswrapper[4767]: I0129 15:14:42.807391 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://dea4eb1d9800844b85bbca49f74990272d70b69b16fee301084f0ca32a79cb1a" gracePeriod=600 Jan 29 15:14:43 crc kubenswrapper[4767]: I0129 15:14:43.332569 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="dea4eb1d9800844b85bbca49f74990272d70b69b16fee301084f0ca32a79cb1a" exitCode=0 Jan 29 15:14:43 crc kubenswrapper[4767]: I0129 15:14:43.332639 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"dea4eb1d9800844b85bbca49f74990272d70b69b16fee301084f0ca32a79cb1a"} Jan 29 15:14:43 crc kubenswrapper[4767]: I0129 15:14:43.332742 4767 scope.go:117] "RemoveContainer" containerID="e4e80bf48c84b6e95f0388dfea83fecc89fce6984a1a924645de1ebde40c9e4c" Jan 29 15:14:44 crc kubenswrapper[4767]: I0129 15:14:44.364534 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"f49c65a1f42f7d0a539525b47266d6427aeb9e7ba92fb408c720097cf89b6158"} Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.174941 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt"] Jan 29 15:15:00 crc kubenswrapper[4767]: E0129 15:15:00.175692 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565418a4-f20c-4841-92fa-d40804d42980" containerName="registry" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.175707 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="565418a4-f20c-4841-92fa-d40804d42980" containerName="registry" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.175830 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="565418a4-f20c-4841-92fa-d40804d42980" containerName="registry" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.176363 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.178657 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.179006 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.184546 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt"] Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.370523 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpw6r\" (UniqueName: \"kubernetes.io/projected/0a83917d-1a8b-470e-8403-d28a2a74a81d-kube-api-access-xpw6r\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.370854 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a83917d-1a8b-470e-8403-d28a2a74a81d-secret-volume\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.370881 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a83917d-1a8b-470e-8403-d28a2a74a81d-config-volume\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.473004 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpw6r\" (UniqueName: \"kubernetes.io/projected/0a83917d-1a8b-470e-8403-d28a2a74a81d-kube-api-access-xpw6r\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.473119 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a83917d-1a8b-470e-8403-d28a2a74a81d-secret-volume\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.473157 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a83917d-1a8b-470e-8403-d28a2a74a81d-config-volume\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.474235 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a83917d-1a8b-470e-8403-d28a2a74a81d-config-volume\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.479345 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a83917d-1a8b-470e-8403-d28a2a74a81d-secret-volume\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.490068 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpw6r\" (UniqueName: \"kubernetes.io/projected/0a83917d-1a8b-470e-8403-d28a2a74a81d-kube-api-access-xpw6r\") pod \"collect-profiles-29494995-mcxqt\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.532332 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:00 crc kubenswrapper[4767]: I0129 15:15:00.707552 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt"] Jan 29 15:15:01 crc kubenswrapper[4767]: I0129 15:15:01.453265 4767 generic.go:334] "Generic (PLEG): container finished" podID="0a83917d-1a8b-470e-8403-d28a2a74a81d" containerID="7dcfed40d25e3788db44ed5e7cda52e787f83fb642353e85eb933f9d04268cbe" exitCode=0 Jan 29 15:15:01 crc kubenswrapper[4767]: I0129 15:15:01.453339 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" event={"ID":"0a83917d-1a8b-470e-8403-d28a2a74a81d","Type":"ContainerDied","Data":"7dcfed40d25e3788db44ed5e7cda52e787f83fb642353e85eb933f9d04268cbe"} Jan 29 15:15:01 crc kubenswrapper[4767]: I0129 15:15:01.453716 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" event={"ID":"0a83917d-1a8b-470e-8403-d28a2a74a81d","Type":"ContainerStarted","Data":"c76e12586b7dc27f1377e63f98269342e21b2a38ea134c276f9cb77023a91ac3"} Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.664883 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.801236 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a83917d-1a8b-470e-8403-d28a2a74a81d-secret-volume\") pod \"0a83917d-1a8b-470e-8403-d28a2a74a81d\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.801357 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a83917d-1a8b-470e-8403-d28a2a74a81d-config-volume\") pod \"0a83917d-1a8b-470e-8403-d28a2a74a81d\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.801407 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpw6r\" (UniqueName: \"kubernetes.io/projected/0a83917d-1a8b-470e-8403-d28a2a74a81d-kube-api-access-xpw6r\") pod \"0a83917d-1a8b-470e-8403-d28a2a74a81d\" (UID: \"0a83917d-1a8b-470e-8403-d28a2a74a81d\") " Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.802173 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a83917d-1a8b-470e-8403-d28a2a74a81d-config-volume" (OuterVolumeSpecName: "config-volume") pod "0a83917d-1a8b-470e-8403-d28a2a74a81d" (UID: "0a83917d-1a8b-470e-8403-d28a2a74a81d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.808889 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a83917d-1a8b-470e-8403-d28a2a74a81d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0a83917d-1a8b-470e-8403-d28a2a74a81d" (UID: "0a83917d-1a8b-470e-8403-d28a2a74a81d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.810005 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a83917d-1a8b-470e-8403-d28a2a74a81d-kube-api-access-xpw6r" (OuterVolumeSpecName: "kube-api-access-xpw6r") pod "0a83917d-1a8b-470e-8403-d28a2a74a81d" (UID: "0a83917d-1a8b-470e-8403-d28a2a74a81d"). InnerVolumeSpecName "kube-api-access-xpw6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.903231 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0a83917d-1a8b-470e-8403-d28a2a74a81d-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.903286 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpw6r\" (UniqueName: \"kubernetes.io/projected/0a83917d-1a8b-470e-8403-d28a2a74a81d-kube-api-access-xpw6r\") on node \"crc\" DevicePath \"\"" Jan 29 15:15:02 crc kubenswrapper[4767]: I0129 15:15:02.903305 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0a83917d-1a8b-470e-8403-d28a2a74a81d-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:15:03 crc kubenswrapper[4767]: I0129 15:15:03.464883 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" event={"ID":"0a83917d-1a8b-470e-8403-d28a2a74a81d","Type":"ContainerDied","Data":"c76e12586b7dc27f1377e63f98269342e21b2a38ea134c276f9cb77023a91ac3"} Jan 29 15:15:03 crc kubenswrapper[4767]: I0129 15:15:03.464946 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c76e12586b7dc27f1377e63f98269342e21b2a38ea134c276f9cb77023a91ac3" Jan 29 15:15:03 crc kubenswrapper[4767]: I0129 15:15:03.465004 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.035218 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-86kzs"] Jan 29 15:15:45 crc kubenswrapper[4767]: E0129 15:15:45.036153 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a83917d-1a8b-470e-8403-d28a2a74a81d" containerName="collect-profiles" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.036170 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a83917d-1a8b-470e-8403-d28a2a74a81d" containerName="collect-profiles" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.036320 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a83917d-1a8b-470e-8403-d28a2a74a81d" containerName="collect-profiles" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.037323 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.043877 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86kzs"] Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.234267 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-utilities\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.234460 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-catalog-content\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.234528 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swgsv\" (UniqueName: \"kubernetes.io/projected/ac128fbf-9b98-49fe-9ab1-a20490c2529d-kube-api-access-swgsv\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.336332 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-utilities\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.336806 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-utilities\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.337172 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-catalog-content\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.336888 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-catalog-content\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.337251 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swgsv\" (UniqueName: \"kubernetes.io/projected/ac128fbf-9b98-49fe-9ab1-a20490c2529d-kube-api-access-swgsv\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.359224 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swgsv\" (UniqueName: \"kubernetes.io/projected/ac128fbf-9b98-49fe-9ab1-a20490c2529d-kube-api-access-swgsv\") pod \"community-operators-86kzs\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.390334 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.640811 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-86kzs"] Jan 29 15:15:45 crc kubenswrapper[4767]: I0129 15:15:45.667534 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86kzs" event={"ID":"ac128fbf-9b98-49fe-9ab1-a20490c2529d","Type":"ContainerStarted","Data":"3b926d27e7b605605e16394c01d5c98b3e783303f0da7c29758fa4a41172a60b"} Jan 29 15:15:46 crc kubenswrapper[4767]: I0129 15:15:46.675127 4767 generic.go:334] "Generic (PLEG): container finished" podID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" containerID="d1f32304868de25c22dc62eb823bdbfa0939ea5d090761aa37af639fb2d9826e" exitCode=0 Jan 29 15:15:46 crc kubenswrapper[4767]: I0129 15:15:46.675186 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86kzs" event={"ID":"ac128fbf-9b98-49fe-9ab1-a20490c2529d","Type":"ContainerDied","Data":"d1f32304868de25c22dc62eb823bdbfa0939ea5d090761aa37af639fb2d9826e"} Jan 29 15:15:46 crc kubenswrapper[4767]: I0129 15:15:46.676626 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:15:46 crc kubenswrapper[4767]: E0129 15:15:46.808810 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:15:46 crc kubenswrapper[4767]: E0129 15:15:46.809077 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swgsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-86kzs_openshift-marketplace(ac128fbf-9b98-49fe-9ab1-a20490c2529d): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:15:46 crc kubenswrapper[4767]: E0129 15:15:46.810325 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:15:47 crc kubenswrapper[4767]: E0129 15:15:47.682236 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.616751 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bv7wp"] Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.618643 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.629080 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bv7wp"] Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.715956 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-catalog-content\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.716217 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-utilities\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.716301 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b68h\" (UniqueName: \"kubernetes.io/projected/371621b9-a052-42ad-ba71-a37a3a064370-kube-api-access-5b68h\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.817629 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-utilities\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.817700 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b68h\" (UniqueName: \"kubernetes.io/projected/371621b9-a052-42ad-ba71-a37a3a064370-kube-api-access-5b68h\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.817772 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-catalog-content\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.818280 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-utilities\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.818337 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-catalog-content\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.853281 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b68h\" (UniqueName: \"kubernetes.io/projected/371621b9-a052-42ad-ba71-a37a3a064370-kube-api-access-5b68h\") pod \"certified-operators-bv7wp\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:15:59 crc kubenswrapper[4767]: I0129 15:15:59.938290 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:16:00 crc kubenswrapper[4767]: W0129 15:16:00.156505 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod371621b9_a052_42ad_ba71_a37a3a064370.slice/crio-f5f660516a00e9a765e881551b69198b3407d0d7bc4602f7b7b963a4b4ef1fc4 WatchSource:0}: Error finding container f5f660516a00e9a765e881551b69198b3407d0d7bc4602f7b7b963a4b4ef1fc4: Status 404 returned error can't find the container with id f5f660516a00e9a765e881551b69198b3407d0d7bc4602f7b7b963a4b4ef1fc4 Jan 29 15:16:00 crc kubenswrapper[4767]: I0129 15:16:00.160031 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bv7wp"] Jan 29 15:16:00 crc kubenswrapper[4767]: I0129 15:16:00.746310 4767 generic.go:334] "Generic (PLEG): container finished" podID="371621b9-a052-42ad-ba71-a37a3a064370" containerID="e48f5f18274bc43ddb52aec03a2c2150e11ba1b57213b4384e1335f3cef16973" exitCode=0 Jan 29 15:16:00 crc kubenswrapper[4767]: I0129 15:16:00.746343 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bv7wp" event={"ID":"371621b9-a052-42ad-ba71-a37a3a064370","Type":"ContainerDied","Data":"e48f5f18274bc43ddb52aec03a2c2150e11ba1b57213b4384e1335f3cef16973"} Jan 29 15:16:00 crc kubenswrapper[4767]: I0129 15:16:00.746365 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bv7wp" event={"ID":"371621b9-a052-42ad-ba71-a37a3a064370","Type":"ContainerStarted","Data":"f5f660516a00e9a765e881551b69198b3407d0d7bc4602f7b7b963a4b4ef1fc4"} Jan 29 15:16:00 crc kubenswrapper[4767]: E0129 15:16:00.908582 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:16:00 crc kubenswrapper[4767]: E0129 15:16:00.908822 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5b68h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bv7wp_openshift-marketplace(371621b9-a052-42ad-ba71-a37a3a064370): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:00 crc kubenswrapper[4767]: E0129 15:16:00.910056 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:16:01 crc kubenswrapper[4767]: E0129 15:16:01.157177 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:16:01 crc kubenswrapper[4767]: E0129 15:16:01.157617 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swgsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-86kzs_openshift-marketplace(ac128fbf-9b98-49fe-9ab1-a20490c2529d): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:01 crc kubenswrapper[4767]: E0129 15:16:01.158983 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:16:01 crc kubenswrapper[4767]: E0129 15:16:01.752588 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:16:12 crc kubenswrapper[4767]: E0129 15:16:12.020745 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:16:13 crc kubenswrapper[4767]: E0129 15:16:13.146553 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:16:13 crc kubenswrapper[4767]: E0129 15:16:13.146680 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5b68h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bv7wp_openshift-marketplace(371621b9-a052-42ad-ba71-a37a3a064370): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:13 crc kubenswrapper[4767]: E0129 15:16:13.148625 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:16:25 crc kubenswrapper[4767]: E0129 15:16:25.022941 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:16:25 crc kubenswrapper[4767]: E0129 15:16:25.160650 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:16:25 crc kubenswrapper[4767]: E0129 15:16:25.160840 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swgsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-86kzs_openshift-marketplace(ac128fbf-9b98-49fe-9ab1-a20490c2529d): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:25 crc kubenswrapper[4767]: E0129 15:16:25.162079 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.351928 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rhqrm"] Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.354545 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.372148 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rhqrm"] Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.452009 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86zw7\" (UniqueName: \"kubernetes.io/projected/1bed5527-b95e-4899-acac-c8191eecb784-kube-api-access-86zw7\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.452199 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-utilities\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.452231 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-catalog-content\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.553409 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-utilities\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.553476 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-catalog-content\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.553509 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86zw7\" (UniqueName: \"kubernetes.io/projected/1bed5527-b95e-4899-acac-c8191eecb784-kube-api-access-86zw7\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.554006 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-catalog-content\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.554044 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-utilities\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.580752 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86zw7\" (UniqueName: \"kubernetes.io/projected/1bed5527-b95e-4899-acac-c8191eecb784-kube-api-access-86zw7\") pod \"redhat-operators-rhqrm\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.727765 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:16:27 crc kubenswrapper[4767]: I0129 15:16:27.918726 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rhqrm"] Jan 29 15:16:28 crc kubenswrapper[4767]: I0129 15:16:28.902850 4767 generic.go:334] "Generic (PLEG): container finished" podID="1bed5527-b95e-4899-acac-c8191eecb784" containerID="7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb" exitCode=0 Jan 29 15:16:28 crc kubenswrapper[4767]: I0129 15:16:28.902923 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rhqrm" event={"ID":"1bed5527-b95e-4899-acac-c8191eecb784","Type":"ContainerDied","Data":"7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb"} Jan 29 15:16:28 crc kubenswrapper[4767]: I0129 15:16:28.903326 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rhqrm" event={"ID":"1bed5527-b95e-4899-acac-c8191eecb784","Type":"ContainerStarted","Data":"4fc06dd4f8eb822a257aa2f9b466b93f9a5504cfa0cb902495d46a97fdb03ca7"} Jan 29 15:16:29 crc kubenswrapper[4767]: E0129 15:16:29.054645 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:16:29 crc kubenswrapper[4767]: E0129 15:16:29.054812 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86zw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rhqrm_openshift-marketplace(1bed5527-b95e-4899-acac-c8191eecb784): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:29 crc kubenswrapper[4767]: E0129 15:16:29.055991 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:16:29 crc kubenswrapper[4767]: E0129 15:16:29.908927 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:16:39 crc kubenswrapper[4767]: E0129 15:16:39.021155 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:16:40 crc kubenswrapper[4767]: E0129 15:16:40.413291 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:16:40 crc kubenswrapper[4767]: E0129 15:16:40.413438 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5b68h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bv7wp_openshift-marketplace(371621b9-a052-42ad-ba71-a37a3a064370): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:40 crc kubenswrapper[4767]: E0129 15:16:40.414961 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:16:44 crc kubenswrapper[4767]: E0129 15:16:44.140777 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:16:44 crc kubenswrapper[4767]: E0129 15:16:44.141316 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86zw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rhqrm_openshift-marketplace(1bed5527-b95e-4899-acac-c8191eecb784): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:44 crc kubenswrapper[4767]: E0129 15:16:44.142479 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.151827 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kb92x"] Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.155015 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.164230 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb92x"] Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.177098 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg8b6\" (UniqueName: \"kubernetes.io/projected/b9fb94e1-edca-4796-a777-d7728207993f-kube-api-access-mg8b6\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.177228 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-catalog-content\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.177260 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-utilities\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.278474 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-catalog-content\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.278989 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-utilities\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.278946 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-catalog-content\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.279084 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg8b6\" (UniqueName: \"kubernetes.io/projected/b9fb94e1-edca-4796-a777-d7728207993f-kube-api-access-mg8b6\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.279305 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-utilities\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.300143 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg8b6\" (UniqueName: \"kubernetes.io/projected/b9fb94e1-edca-4796-a777-d7728207993f-kube-api-access-mg8b6\") pod \"redhat-marketplace-kb92x\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.474161 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.672992 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb92x"] Jan 29 15:16:46 crc kubenswrapper[4767]: W0129 15:16:46.680193 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9fb94e1_edca_4796_a777_d7728207993f.slice/crio-c03089076542a7f969a703dc077e0db542dd6e30102f5abb35d5b3003b532f86 WatchSource:0}: Error finding container c03089076542a7f969a703dc077e0db542dd6e30102f5abb35d5b3003b532f86: Status 404 returned error can't find the container with id c03089076542a7f969a703dc077e0db542dd6e30102f5abb35d5b3003b532f86 Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.996131 4767 generic.go:334] "Generic (PLEG): container finished" podID="b9fb94e1-edca-4796-a777-d7728207993f" containerID="be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689" exitCode=0 Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.996192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb92x" event={"ID":"b9fb94e1-edca-4796-a777-d7728207993f","Type":"ContainerDied","Data":"be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689"} Jan 29 15:16:46 crc kubenswrapper[4767]: I0129 15:16:46.996220 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb92x" event={"ID":"b9fb94e1-edca-4796-a777-d7728207993f","Type":"ContainerStarted","Data":"c03089076542a7f969a703dc077e0db542dd6e30102f5abb35d5b3003b532f86"} Jan 29 15:16:47 crc kubenswrapper[4767]: E0129 15:16:47.116953 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:16:47 crc kubenswrapper[4767]: E0129 15:16:47.117146 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mg8b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kb92x_openshift-marketplace(b9fb94e1-edca-4796-a777-d7728207993f): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:47 crc kubenswrapper[4767]: E0129 15:16:47.118363 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" Jan 29 15:16:48 crc kubenswrapper[4767]: E0129 15:16:48.002019 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" Jan 29 15:16:53 crc kubenswrapper[4767]: E0129 15:16:53.021098 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:16:55 crc kubenswrapper[4767]: E0129 15:16:55.023464 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:16:56 crc kubenswrapper[4767]: E0129 15:16:56.020102 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:16:59 crc kubenswrapper[4767]: E0129 15:16:59.238632 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:16:59 crc kubenswrapper[4767]: E0129 15:16:59.239116 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mg8b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kb92x_openshift-marketplace(b9fb94e1-edca-4796-a777-d7728207993f): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:16:59 crc kubenswrapper[4767]: E0129 15:16:59.240322 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" Jan 29 15:17:05 crc kubenswrapper[4767]: E0129 15:17:05.022178 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:17:08 crc kubenswrapper[4767]: E0129 15:17:08.020168 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:17:08 crc kubenswrapper[4767]: E0129 15:17:08.147576 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:17:08 crc kubenswrapper[4767]: E0129 15:17:08.147736 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86zw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-rhqrm_openshift-marketplace(1bed5527-b95e-4899-acac-c8191eecb784): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:17:08 crc kubenswrapper[4767]: E0129 15:17:08.148940 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:17:11 crc kubenswrapper[4767]: E0129 15:17:11.020839 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" Jan 29 15:17:12 crc kubenswrapper[4767]: I0129 15:17:12.806039 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:17:12 crc kubenswrapper[4767]: I0129 15:17:12.806604 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:17:20 crc kubenswrapper[4767]: E0129 15:17:20.205622 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:17:20 crc kubenswrapper[4767]: E0129 15:17:20.206254 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swgsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-86kzs_openshift-marketplace(ac128fbf-9b98-49fe-9ab1-a20490c2529d): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:17:20 crc kubenswrapper[4767]: E0129 15:17:20.208111 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:17:22 crc kubenswrapper[4767]: E0129 15:17:22.021967 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:17:22 crc kubenswrapper[4767]: E0129 15:17:22.160252 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:17:22 crc kubenswrapper[4767]: E0129 15:17:22.160408 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5b68h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bv7wp_openshift-marketplace(371621b9-a052-42ad-ba71-a37a3a064370): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:17:22 crc kubenswrapper[4767]: E0129 15:17:22.161730 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:17:22 crc kubenswrapper[4767]: E0129 15:17:22.170800 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:17:22 crc kubenswrapper[4767]: E0129 15:17:22.170939 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mg8b6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kb92x_openshift-marketplace(b9fb94e1-edca-4796-a777-d7728207993f): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:17:22 crc kubenswrapper[4767]: E0129 15:17:22.172118 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" Jan 29 15:17:31 crc kubenswrapper[4767]: E0129 15:17:31.020092 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:17:33 crc kubenswrapper[4767]: E0129 15:17:33.022240 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:17:34 crc kubenswrapper[4767]: E0129 15:17:34.019946 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:17:36 crc kubenswrapper[4767]: E0129 15:17:36.020204 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" Jan 29 15:17:42 crc kubenswrapper[4767]: E0129 15:17:42.020557 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:17:42 crc kubenswrapper[4767]: I0129 15:17:42.805675 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:17:42 crc kubenswrapper[4767]: I0129 15:17:42.805730 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:17:47 crc kubenswrapper[4767]: E0129 15:17:47.021053 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" Jan 29 15:17:48 crc kubenswrapper[4767]: E0129 15:17:48.021043 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:17:51 crc kubenswrapper[4767]: E0129 15:17:51.021499 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" Jan 29 15:17:57 crc kubenswrapper[4767]: E0129 15:17:57.019857 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:18:03 crc kubenswrapper[4767]: E0129 15:18:03.020758 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:18:03 crc kubenswrapper[4767]: I0129 15:18:03.433512 4767 generic.go:334] "Generic (PLEG): container finished" podID="1bed5527-b95e-4899-acac-c8191eecb784" containerID="cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea" exitCode=0 Jan 29 15:18:03 crc kubenswrapper[4767]: I0129 15:18:03.433568 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rhqrm" event={"ID":"1bed5527-b95e-4899-acac-c8191eecb784","Type":"ContainerDied","Data":"cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea"} Jan 29 15:18:04 crc kubenswrapper[4767]: I0129 15:18:04.440677 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rhqrm" event={"ID":"1bed5527-b95e-4899-acac-c8191eecb784","Type":"ContainerStarted","Data":"b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d"} Jan 29 15:18:04 crc kubenswrapper[4767]: I0129 15:18:04.464699 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rhqrm" podStartSLOduration=2.136871899 podStartE2EDuration="1m37.464682423s" podCreationTimestamp="2026-01-29 15:16:27 +0000 UTC" firstStartedPulling="2026-01-29 15:16:28.905058503 +0000 UTC m=+944.715375542" lastFinishedPulling="2026-01-29 15:18:04.232869007 +0000 UTC m=+1040.043186066" observedRunningTime="2026-01-29 15:18:04.460060287 +0000 UTC m=+1040.270377346" watchObservedRunningTime="2026-01-29 15:18:04.464682423 +0000 UTC m=+1040.274999462" Jan 29 15:18:07 crc kubenswrapper[4767]: I0129 15:18:07.459955 4767 generic.go:334] "Generic (PLEG): container finished" podID="b9fb94e1-edca-4796-a777-d7728207993f" containerID="86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5" exitCode=0 Jan 29 15:18:07 crc kubenswrapper[4767]: I0129 15:18:07.460025 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb92x" event={"ID":"b9fb94e1-edca-4796-a777-d7728207993f","Type":"ContainerDied","Data":"86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5"} Jan 29 15:18:07 crc kubenswrapper[4767]: I0129 15:18:07.728483 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:18:07 crc kubenswrapper[4767]: I0129 15:18:07.728544 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:18:08 crc kubenswrapper[4767]: I0129 15:18:08.466967 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb92x" event={"ID":"b9fb94e1-edca-4796-a777-d7728207993f","Type":"ContainerStarted","Data":"199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9"} Jan 29 15:18:08 crc kubenswrapper[4767]: I0129 15:18:08.491376 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kb92x" podStartSLOduration=1.633278766 podStartE2EDuration="1m22.491358551s" podCreationTimestamp="2026-01-29 15:16:46 +0000 UTC" firstStartedPulling="2026-01-29 15:16:46.997525878 +0000 UTC m=+962.807842917" lastFinishedPulling="2026-01-29 15:18:07.855605663 +0000 UTC m=+1043.665922702" observedRunningTime="2026-01-29 15:18:08.486989632 +0000 UTC m=+1044.297306671" watchObservedRunningTime="2026-01-29 15:18:08.491358551 +0000 UTC m=+1044.301675590" Jan 29 15:18:08 crc kubenswrapper[4767]: I0129 15:18:08.772205 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="registry-server" probeResult="failure" output=< Jan 29 15:18:08 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:18:08 crc kubenswrapper[4767]: > Jan 29 15:18:12 crc kubenswrapper[4767]: E0129 15:18:12.021323 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:18:12 crc kubenswrapper[4767]: I0129 15:18:12.805380 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:18:12 crc kubenswrapper[4767]: I0129 15:18:12.805440 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:18:12 crc kubenswrapper[4767]: I0129 15:18:12.805494 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:18:12 crc kubenswrapper[4767]: I0129 15:18:12.806045 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f49c65a1f42f7d0a539525b47266d6427aeb9e7ba92fb408c720097cf89b6158"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:18:12 crc kubenswrapper[4767]: I0129 15:18:12.806118 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://f49c65a1f42f7d0a539525b47266d6427aeb9e7ba92fb408c720097cf89b6158" gracePeriod=600 Jan 29 15:18:13 crc kubenswrapper[4767]: I0129 15:18:13.499111 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="f49c65a1f42f7d0a539525b47266d6427aeb9e7ba92fb408c720097cf89b6158" exitCode=0 Jan 29 15:18:13 crc kubenswrapper[4767]: I0129 15:18:13.499132 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"f49c65a1f42f7d0a539525b47266d6427aeb9e7ba92fb408c720097cf89b6158"} Jan 29 15:18:13 crc kubenswrapper[4767]: I0129 15:18:13.499746 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"c74750ca0dbd8fc159be6c3f13ddce80420e211a2a5224b69e150664c60acb0e"} Jan 29 15:18:13 crc kubenswrapper[4767]: I0129 15:18:13.499777 4767 scope.go:117] "RemoveContainer" containerID="dea4eb1d9800844b85bbca49f74990272d70b69b16fee301084f0ca32a79cb1a" Jan 29 15:18:16 crc kubenswrapper[4767]: I0129 15:18:16.475091 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:18:16 crc kubenswrapper[4767]: I0129 15:18:16.477102 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:18:16 crc kubenswrapper[4767]: I0129 15:18:16.529110 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:18:16 crc kubenswrapper[4767]: I0129 15:18:16.571659 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:18:17 crc kubenswrapper[4767]: I0129 15:18:17.404539 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb92x"] Jan 29 15:18:17 crc kubenswrapper[4767]: I0129 15:18:17.789061 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:18:17 crc kubenswrapper[4767]: I0129 15:18:17.842675 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:18:18 crc kubenswrapper[4767]: E0129 15:18:18.020936 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:18:18 crc kubenswrapper[4767]: I0129 15:18:18.529429 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kb92x" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="registry-server" containerID="cri-o://199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9" gracePeriod=2 Jan 29 15:18:18 crc kubenswrapper[4767]: I0129 15:18:18.882954 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.083075 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg8b6\" (UniqueName: \"kubernetes.io/projected/b9fb94e1-edca-4796-a777-d7728207993f-kube-api-access-mg8b6\") pod \"b9fb94e1-edca-4796-a777-d7728207993f\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.083173 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-utilities\") pod \"b9fb94e1-edca-4796-a777-d7728207993f\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.083339 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-catalog-content\") pod \"b9fb94e1-edca-4796-a777-d7728207993f\" (UID: \"b9fb94e1-edca-4796-a777-d7728207993f\") " Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.084673 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-utilities" (OuterVolumeSpecName: "utilities") pod "b9fb94e1-edca-4796-a777-d7728207993f" (UID: "b9fb94e1-edca-4796-a777-d7728207993f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.090701 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9fb94e1-edca-4796-a777-d7728207993f-kube-api-access-mg8b6" (OuterVolumeSpecName: "kube-api-access-mg8b6") pod "b9fb94e1-edca-4796-a777-d7728207993f" (UID: "b9fb94e1-edca-4796-a777-d7728207993f"). InnerVolumeSpecName "kube-api-access-mg8b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.109993 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9fb94e1-edca-4796-a777-d7728207993f" (UID: "b9fb94e1-edca-4796-a777-d7728207993f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.185087 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.185342 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg8b6\" (UniqueName: \"kubernetes.io/projected/b9fb94e1-edca-4796-a777-d7728207993f-kube-api-access-mg8b6\") on node \"crc\" DevicePath \"\"" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.185420 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9fb94e1-edca-4796-a777-d7728207993f-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.537036 4767 generic.go:334] "Generic (PLEG): container finished" podID="b9fb94e1-edca-4796-a777-d7728207993f" containerID="199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9" exitCode=0 Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.537087 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb92x" event={"ID":"b9fb94e1-edca-4796-a777-d7728207993f","Type":"ContainerDied","Data":"199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9"} Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.537112 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb92x" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.537139 4767 scope.go:117] "RemoveContainer" containerID="199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.537126 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb92x" event={"ID":"b9fb94e1-edca-4796-a777-d7728207993f","Type":"ContainerDied","Data":"c03089076542a7f969a703dc077e0db542dd6e30102f5abb35d5b3003b532f86"} Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.566017 4767 scope.go:117] "RemoveContainer" containerID="86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.566777 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb92x"] Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.572825 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb92x"] Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.596931 4767 scope.go:117] "RemoveContainer" containerID="be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.612316 4767 scope.go:117] "RemoveContainer" containerID="199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9" Jan 29 15:18:19 crc kubenswrapper[4767]: E0129 15:18:19.612751 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9\": container with ID starting with 199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9 not found: ID does not exist" containerID="199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.612932 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9"} err="failed to get container status \"199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9\": rpc error: code = NotFound desc = could not find container \"199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9\": container with ID starting with 199bcbe32a67204df80852646d54e5c16890c0db72ae89d3dac209506c2b5de9 not found: ID does not exist" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.613085 4767 scope.go:117] "RemoveContainer" containerID="86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5" Jan 29 15:18:19 crc kubenswrapper[4767]: E0129 15:18:19.613527 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5\": container with ID starting with 86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5 not found: ID does not exist" containerID="86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.613589 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5"} err="failed to get container status \"86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5\": rpc error: code = NotFound desc = could not find container \"86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5\": container with ID starting with 86b9f819795b3469a8041b7767d1f7ae5c18e517173312575681af497c8320d5 not found: ID does not exist" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.613622 4767 scope.go:117] "RemoveContainer" containerID="be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689" Jan 29 15:18:19 crc kubenswrapper[4767]: E0129 15:18:19.614007 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689\": container with ID starting with be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689 not found: ID does not exist" containerID="be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689" Jan 29 15:18:19 crc kubenswrapper[4767]: I0129 15:18:19.614040 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689"} err="failed to get container status \"be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689\": rpc error: code = NotFound desc = could not find container \"be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689\": container with ID starting with be7558ee90c28bfa4fb24b9e862456827776596670b05b094073463069286689 not found: ID does not exist" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.199725 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rhqrm"] Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.200022 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rhqrm" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="registry-server" containerID="cri-o://b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d" gracePeriod=2 Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.519101 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.545110 4767 generic.go:334] "Generic (PLEG): container finished" podID="1bed5527-b95e-4899-acac-c8191eecb784" containerID="b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d" exitCode=0 Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.545159 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rhqrm" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.545222 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rhqrm" event={"ID":"1bed5527-b95e-4899-acac-c8191eecb784","Type":"ContainerDied","Data":"b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d"} Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.545294 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rhqrm" event={"ID":"1bed5527-b95e-4899-acac-c8191eecb784","Type":"ContainerDied","Data":"4fc06dd4f8eb822a257aa2f9b466b93f9a5504cfa0cb902495d46a97fdb03ca7"} Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.545316 4767 scope.go:117] "RemoveContainer" containerID="b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.558062 4767 scope.go:117] "RemoveContainer" containerID="cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.574631 4767 scope.go:117] "RemoveContainer" containerID="7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.592252 4767 scope.go:117] "RemoveContainer" containerID="b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d" Jan 29 15:18:20 crc kubenswrapper[4767]: E0129 15:18:20.593001 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d\": container with ID starting with b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d not found: ID does not exist" containerID="b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.593064 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d"} err="failed to get container status \"b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d\": rpc error: code = NotFound desc = could not find container \"b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d\": container with ID starting with b82328bd1a5db8d438375dd4b5a0abb1bef162657f536affdca806b022ec700d not found: ID does not exist" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.593099 4767 scope.go:117] "RemoveContainer" containerID="cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea" Jan 29 15:18:20 crc kubenswrapper[4767]: E0129 15:18:20.593565 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea\": container with ID starting with cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea not found: ID does not exist" containerID="cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.593612 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea"} err="failed to get container status \"cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea\": rpc error: code = NotFound desc = could not find container \"cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea\": container with ID starting with cf7a68f4691fa36bcf478a2d013f87557498911b0b98cb6986033b92464618ea not found: ID does not exist" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.593646 4767 scope.go:117] "RemoveContainer" containerID="7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb" Jan 29 15:18:20 crc kubenswrapper[4767]: E0129 15:18:20.594106 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb\": container with ID starting with 7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb not found: ID does not exist" containerID="7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.594136 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb"} err="failed to get container status \"7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb\": rpc error: code = NotFound desc = could not find container \"7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb\": container with ID starting with 7e69e4df0b56d4bbf3abf077b60203f3296818617da89d71b40a7820f0b558bb not found: ID does not exist" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.705723 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-utilities\") pod \"1bed5527-b95e-4899-acac-c8191eecb784\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.705779 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-catalog-content\") pod \"1bed5527-b95e-4899-acac-c8191eecb784\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.705833 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86zw7\" (UniqueName: \"kubernetes.io/projected/1bed5527-b95e-4899-acac-c8191eecb784-kube-api-access-86zw7\") pod \"1bed5527-b95e-4899-acac-c8191eecb784\" (UID: \"1bed5527-b95e-4899-acac-c8191eecb784\") " Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.706696 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-utilities" (OuterVolumeSpecName: "utilities") pod "1bed5527-b95e-4899-acac-c8191eecb784" (UID: "1bed5527-b95e-4899-acac-c8191eecb784"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.714262 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bed5527-b95e-4899-acac-c8191eecb784-kube-api-access-86zw7" (OuterVolumeSpecName: "kube-api-access-86zw7") pod "1bed5527-b95e-4899-acac-c8191eecb784" (UID: "1bed5527-b95e-4899-acac-c8191eecb784"). InnerVolumeSpecName "kube-api-access-86zw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.806888 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86zw7\" (UniqueName: \"kubernetes.io/projected/1bed5527-b95e-4899-acac-c8191eecb784-kube-api-access-86zw7\") on node \"crc\" DevicePath \"\"" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.806955 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.823066 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1bed5527-b95e-4899-acac-c8191eecb784" (UID: "1bed5527-b95e-4899-acac-c8191eecb784"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.872054 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rhqrm"] Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.875633 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rhqrm"] Jan 29 15:18:20 crc kubenswrapper[4767]: I0129 15:18:20.908261 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1bed5527-b95e-4899-acac-c8191eecb784-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:18:21 crc kubenswrapper[4767]: I0129 15:18:21.025547 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bed5527-b95e-4899-acac-c8191eecb784" path="/var/lib/kubelet/pods/1bed5527-b95e-4899-acac-c8191eecb784/volumes" Jan 29 15:18:21 crc kubenswrapper[4767]: I0129 15:18:21.026222 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9fb94e1-edca-4796-a777-d7728207993f" path="/var/lib/kubelet/pods/b9fb94e1-edca-4796-a777-d7728207993f/volumes" Jan 29 15:18:25 crc kubenswrapper[4767]: E0129 15:18:25.025016 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:18:29 crc kubenswrapper[4767]: E0129 15:18:29.021226 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:18:39 crc kubenswrapper[4767]: E0129 15:18:39.021631 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:18:41 crc kubenswrapper[4767]: E0129 15:18:41.021456 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:18:54 crc kubenswrapper[4767]: E0129 15:18:54.171669 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:18:54 crc kubenswrapper[4767]: E0129 15:18:54.172560 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-swgsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-86kzs_openshift-marketplace(ac128fbf-9b98-49fe-9ab1-a20490c2529d): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:18:54 crc kubenswrapper[4767]: E0129 15:18:54.173790 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:18:55 crc kubenswrapper[4767]: E0129 15:18:55.171853 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:18:55 crc kubenswrapper[4767]: E0129 15:18:55.172165 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5b68h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-bv7wp_openshift-marketplace(371621b9-a052-42ad-ba71-a37a3a064370): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:18:55 crc kubenswrapper[4767]: E0129 15:18:55.173444 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:19:06 crc kubenswrapper[4767]: E0129 15:19:06.023932 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:19:09 crc kubenswrapper[4767]: E0129 15:19:09.020721 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:19:18 crc kubenswrapper[4767]: E0129 15:19:18.020869 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:19:21 crc kubenswrapper[4767]: E0129 15:19:21.025125 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:19:32 crc kubenswrapper[4767]: E0129 15:19:32.021048 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:19:33 crc kubenswrapper[4767]: E0129 15:19:33.021192 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:19:45 crc kubenswrapper[4767]: E0129 15:19:45.024676 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:19:46 crc kubenswrapper[4767]: E0129 15:19:46.020227 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:19:57 crc kubenswrapper[4767]: E0129 15:19:57.022793 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:19:57 crc kubenswrapper[4767]: E0129 15:19:57.023264 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:20:09 crc kubenswrapper[4767]: E0129 15:20:09.022310 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:20:10 crc kubenswrapper[4767]: E0129 15:20:10.020618 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.660312 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tg5zk"] Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.661718 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-controller" containerID="cri-o://e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.661838 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="sbdb" containerID="cri-o://f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.661925 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="nbdb" containerID="cri-o://fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.661946 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.661841 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="northd" containerID="cri-o://0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.662044 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-acl-logging" containerID="cri-o://de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.662081 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-node" containerID="cri-o://8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.705307 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" containerID="cri-o://d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" gracePeriod=30 Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.945127 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/3.log" Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.948590 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovn-acl-logging/0.log" Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.949725 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovn-controller/0.log" Jan 29 15:20:17 crc kubenswrapper[4767]: I0129 15:20:17.950240 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006454 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kqqcl"] Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006693 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006708 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006719 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="extract-content" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006727 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="extract-content" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006735 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006742 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006752 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006759 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006769 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="registry-server" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006777 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="registry-server" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006790 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kubecfg-setup" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006799 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kubecfg-setup" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006807 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="registry-server" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006815 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="registry-server" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006824 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="sbdb" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006831 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="sbdb" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006842 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006849 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006856 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="extract-utilities" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006863 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="extract-utilities" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006873 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="northd" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006881 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="northd" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006908 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="extract-utilities" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006916 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="extract-utilities" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006926 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-node" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006933 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-node" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006943 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006950 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006965 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="extract-content" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006972 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="extract-content" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.006984 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-acl-logging" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.006990 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-acl-logging" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.007001 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007007 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.007017 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="nbdb" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007024 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="nbdb" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007132 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007145 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007154 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007163 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007171 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007180 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-node" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007189 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bed5527-b95e-4899-acac-c8191eecb784" containerName="registry-server" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007201 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="nbdb" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007214 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovn-acl-logging" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007225 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="kube-rbac-proxy-ovn-metrics" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007234 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9fb94e1-edca-4796-a777-d7728207993f" containerName="registry-server" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007244 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="northd" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007254 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="sbdb" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.007358 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007368 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.007480 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="633e55cd-6740-4d13-84ef-023b691c0428" containerName="ovnkube-controller" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.009405 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032176 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-systemd\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032261 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-script-lib\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032302 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/633e55cd-6740-4d13-84ef-023b691c0428-ovn-node-metrics-cert\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032322 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-node-log\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032358 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-kubelet\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032394 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw776\" (UniqueName: \"kubernetes.io/projected/633e55cd-6740-4d13-84ef-023b691c0428-kube-api-access-dw776\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032446 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-env-overrides\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032479 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-netd\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032507 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-config\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032529 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-ovn-kubernetes\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032551 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-bin\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032540 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032581 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-log-socket\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032592 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032608 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-etc-openvswitch\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032626 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032670 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032674 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-var-lib-openvswitch\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032727 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-slash\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032759 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-netns\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032787 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-systemd-units\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032816 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-openvswitch\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032695 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032874 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032707 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-node-log" (OuterVolumeSpecName: "node-log") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032792 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-log-socket" (OuterVolumeSpecName: "log-socket") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032944 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032816 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032844 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-slash" (OuterVolumeSpecName: "host-slash") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032974 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.032849 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-ovn\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033002 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033016 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033055 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033111 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-var-lib-cni-networks-ovn-kubernetes\") pod \"633e55cd-6740-4d13-84ef-023b691c0428\" (UID: \"633e55cd-6740-4d13-84ef-023b691c0428\") " Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033205 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033215 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033744 4767 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033763 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033775 4767 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033789 4767 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033806 4767 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033818 4767 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-log-socket\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033829 4767 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033840 4767 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033851 4767 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-slash\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033865 4767 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033876 4767 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033888 4767 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033923 4767 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033936 4767 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033949 4767 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/633e55cd-6740-4d13-84ef-023b691c0428-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033962 4767 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-node-log\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.033974 4767 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.039649 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/633e55cd-6740-4d13-84ef-023b691c0428-kube-api-access-dw776" (OuterVolumeSpecName: "kube-api-access-dw776") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "kube-api-access-dw776". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.041516 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/633e55cd-6740-4d13-84ef-023b691c0428-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.048356 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "633e55cd-6740-4d13-84ef-023b691c0428" (UID: "633e55cd-6740-4d13-84ef-023b691c0428"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.135423 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-etc-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.135862 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-cni-bin\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.136031 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-ovn\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.136204 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.136350 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.136489 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovnkube-script-lib\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.136618 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-var-lib-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.136798 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-cni-netd\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.136994 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovnkube-config\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.137136 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-systemd-units\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.137299 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfjfp\" (UniqueName: \"kubernetes.io/projected/faaa8fe4-f480-4031-b6b7-c883d62209ee-kube-api-access-vfjfp\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.137450 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-systemd\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.137593 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-run-netns\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.137771 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-node-log\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138074 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-kubelet\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138123 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovn-node-metrics-cert\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138145 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-log-socket\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138163 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-slash\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138186 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-env-overrides\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138206 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-run-ovn-kubernetes\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138330 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw776\" (UniqueName: \"kubernetes.io/projected/633e55cd-6740-4d13-84ef-023b691c0428-kube-api-access-dw776\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138363 4767 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/633e55cd-6740-4d13-84ef-023b691c0428-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.138381 4767 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/633e55cd-6740-4d13-84ef-023b691c0428-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.219305 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovnkube-controller/3.log" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222028 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovn-acl-logging/0.log" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222483 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tg5zk_633e55cd-6740-4d13-84ef-023b691c0428/ovn-controller/0.log" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222790 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" exitCode=0 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222846 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222817 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" exitCode=0 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222884 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" exitCode=0 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222917 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" exitCode=0 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222928 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" exitCode=0 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222938 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" exitCode=0 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222947 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" exitCode=143 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222956 4767 generic.go:334] "Generic (PLEG): container finished" podID="633e55cd-6740-4d13-84ef-023b691c0428" containerID="e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" exitCode=143 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.222843 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223016 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223027 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223037 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223046 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223057 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223070 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223080 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223087 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223093 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223098 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223104 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223110 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223115 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223120 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223127 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223135 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223140 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223145 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223150 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223154 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223159 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223164 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223169 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223176 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223180 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223187 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223194 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223200 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223205 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223210 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223216 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223221 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223227 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223232 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223237 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223242 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223248 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tg5zk" event={"ID":"633e55cd-6740-4d13-84ef-023b691c0428","Type":"ContainerDied","Data":"6660b45896ac7dd6baa4aa6e19ab9a27d30bbe942e7ba0aa3826282c494afe7b"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223256 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223262 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223267 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223272 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223277 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223282 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223287 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223292 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223297 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223302 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.223314 4767 scope.go:117] "RemoveContainer" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.225620 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/2.log" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.226088 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/1.log" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.226116 4767 generic.go:334] "Generic (PLEG): container finished" podID="4615231a-e157-430e-9ecc-97c3fd7701bb" containerID="f1630240852f2539ae4c87bd3e8b6e0a401724521e6bb6888ffe71ccb369f8f3" exitCode=2 Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.226134 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerDied","Data":"f1630240852f2539ae4c87bd3e8b6e0a401724521e6bb6888ffe71ccb369f8f3"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.226146 4767 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b"} Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.226419 4767 scope.go:117] "RemoveContainer" containerID="f1630240852f2539ae4c87bd3e8b6e0a401724521e6bb6888ffe71ccb369f8f3" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239690 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-cni-bin\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239753 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-ovn\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239777 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239795 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239813 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovnkube-script-lib\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239917 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239927 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-ovn\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239937 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239978 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-var-lib-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.239958 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-var-lib-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240078 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-cni-netd\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240121 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovnkube-config\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240152 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-cni-netd\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240168 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-systemd-units\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240214 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-systemd-units\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240223 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfjfp\" (UniqueName: \"kubernetes.io/projected/faaa8fe4-f480-4031-b6b7-c883d62209ee-kube-api-access-vfjfp\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240300 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-systemd\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240339 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-run-netns\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240397 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-node-log\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240424 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-run-systemd\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240465 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-kubelet\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240509 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-kubelet\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240437 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-run-netns\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240542 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovn-node-metrics-cert\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240591 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-slash\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240580 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-node-log\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240635 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-log-socket\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240667 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-slash\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240677 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-env-overrides\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240699 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovnkube-script-lib\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240672 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-log-socket\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240732 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-run-ovn-kubernetes\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240764 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-etc-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240792 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-run-ovn-kubernetes\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.240830 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-etc-openvswitch\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.241324 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-env-overrides\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.241376 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovnkube-config\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.241740 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/faaa8fe4-f480-4031-b6b7-c883d62209ee-host-cni-bin\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.245570 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/faaa8fe4-f480-4031-b6b7-c883d62209ee-ovn-node-metrics-cert\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.256151 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.268461 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfjfp\" (UniqueName: \"kubernetes.io/projected/faaa8fe4-f480-4031-b6b7-c883d62209ee-kube-api-access-vfjfp\") pod \"ovnkube-node-kqqcl\" (UID: \"faaa8fe4-f480-4031-b6b7-c883d62209ee\") " pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.275400 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tg5zk"] Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.275624 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tg5zk"] Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.291760 4767 scope.go:117] "RemoveContainer" containerID="f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.316288 4767 scope.go:117] "RemoveContainer" containerID="fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.321353 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.339087 4767 scope.go:117] "RemoveContainer" containerID="0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.352159 4767 scope.go:117] "RemoveContainer" containerID="fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.379134 4767 scope.go:117] "RemoveContainer" containerID="8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.392150 4767 scope.go:117] "RemoveContainer" containerID="de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.407665 4767 scope.go:117] "RemoveContainer" containerID="e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.420944 4767 scope.go:117] "RemoveContainer" containerID="a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.436504 4767 scope.go:117] "RemoveContainer" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.436977 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": container with ID starting with d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154 not found: ID does not exist" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.437007 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} err="failed to get container status \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": rpc error: code = NotFound desc = could not find container \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": container with ID starting with d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.437031 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.437301 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": container with ID starting with 1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6 not found: ID does not exist" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.437319 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} err="failed to get container status \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": rpc error: code = NotFound desc = could not find container \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": container with ID starting with 1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.437332 4767 scope.go:117] "RemoveContainer" containerID="f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.437565 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": container with ID starting with f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5 not found: ID does not exist" containerID="f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.437667 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} err="failed to get container status \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": rpc error: code = NotFound desc = could not find container \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": container with ID starting with f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.437744 4767 scope.go:117] "RemoveContainer" containerID="fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.438106 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": container with ID starting with fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80 not found: ID does not exist" containerID="fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.438126 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} err="failed to get container status \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": rpc error: code = NotFound desc = could not find container \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": container with ID starting with fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.438141 4767 scope.go:117] "RemoveContainer" containerID="0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.438365 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": container with ID starting with 0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e not found: ID does not exist" containerID="0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.438437 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} err="failed to get container status \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": rpc error: code = NotFound desc = could not find container \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": container with ID starting with 0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.438505 4767 scope.go:117] "RemoveContainer" containerID="fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.438810 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": container with ID starting with fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570 not found: ID does not exist" containerID="fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.438847 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} err="failed to get container status \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": rpc error: code = NotFound desc = could not find container \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": container with ID starting with fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.438874 4767 scope.go:117] "RemoveContainer" containerID="8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.439182 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": container with ID starting with 8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba not found: ID does not exist" containerID="8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.439264 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} err="failed to get container status \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": rpc error: code = NotFound desc = could not find container \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": container with ID starting with 8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.439324 4767 scope.go:117] "RemoveContainer" containerID="de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.439652 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": container with ID starting with de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c not found: ID does not exist" containerID="de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.439679 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} err="failed to get container status \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": rpc error: code = NotFound desc = could not find container \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": container with ID starting with de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.439694 4767 scope.go:117] "RemoveContainer" containerID="e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.439981 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": container with ID starting with e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e not found: ID does not exist" containerID="e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.440071 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} err="failed to get container status \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": rpc error: code = NotFound desc = could not find container \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": container with ID starting with e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.440139 4767 scope.go:117] "RemoveContainer" containerID="a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad" Jan 29 15:20:18 crc kubenswrapper[4767]: E0129 15:20:18.440401 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": container with ID starting with a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad not found: ID does not exist" containerID="a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.440473 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} err="failed to get container status \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": rpc error: code = NotFound desc = could not find container \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": container with ID starting with a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.440531 4767 scope.go:117] "RemoveContainer" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.440799 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} err="failed to get container status \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": rpc error: code = NotFound desc = could not find container \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": container with ID starting with d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.440818 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.441035 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} err="failed to get container status \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": rpc error: code = NotFound desc = could not find container \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": container with ID starting with 1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.441103 4767 scope.go:117] "RemoveContainer" containerID="f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.442228 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} err="failed to get container status \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": rpc error: code = NotFound desc = could not find container \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": container with ID starting with f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.442251 4767 scope.go:117] "RemoveContainer" containerID="fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.442472 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} err="failed to get container status \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": rpc error: code = NotFound desc = could not find container \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": container with ID starting with fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.442519 4767 scope.go:117] "RemoveContainer" containerID="0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.442769 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} err="failed to get container status \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": rpc error: code = NotFound desc = could not find container \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": container with ID starting with 0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.442802 4767 scope.go:117] "RemoveContainer" containerID="fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443034 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} err="failed to get container status \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": rpc error: code = NotFound desc = could not find container \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": container with ID starting with fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443066 4767 scope.go:117] "RemoveContainer" containerID="8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443248 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} err="failed to get container status \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": rpc error: code = NotFound desc = could not find container \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": container with ID starting with 8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443316 4767 scope.go:117] "RemoveContainer" containerID="de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443552 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} err="failed to get container status \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": rpc error: code = NotFound desc = could not find container \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": container with ID starting with de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443582 4767 scope.go:117] "RemoveContainer" containerID="e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443888 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} err="failed to get container status \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": rpc error: code = NotFound desc = could not find container \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": container with ID starting with e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.443929 4767 scope.go:117] "RemoveContainer" containerID="a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.444145 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} err="failed to get container status \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": rpc error: code = NotFound desc = could not find container \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": container with ID starting with a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.444276 4767 scope.go:117] "RemoveContainer" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.444604 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} err="failed to get container status \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": rpc error: code = NotFound desc = could not find container \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": container with ID starting with d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.444655 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.445307 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} err="failed to get container status \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": rpc error: code = NotFound desc = could not find container \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": container with ID starting with 1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.445381 4767 scope.go:117] "RemoveContainer" containerID="f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.445717 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} err="failed to get container status \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": rpc error: code = NotFound desc = could not find container \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": container with ID starting with f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.445740 4767 scope.go:117] "RemoveContainer" containerID="fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.446024 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} err="failed to get container status \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": rpc error: code = NotFound desc = could not find container \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": container with ID starting with fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.446103 4767 scope.go:117] "RemoveContainer" containerID="0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.446381 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} err="failed to get container status \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": rpc error: code = NotFound desc = could not find container \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": container with ID starting with 0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.446429 4767 scope.go:117] "RemoveContainer" containerID="fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.446648 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} err="failed to get container status \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": rpc error: code = NotFound desc = could not find container \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": container with ID starting with fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.446718 4767 scope.go:117] "RemoveContainer" containerID="8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.447020 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} err="failed to get container status \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": rpc error: code = NotFound desc = could not find container \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": container with ID starting with 8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.447065 4767 scope.go:117] "RemoveContainer" containerID="de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.447272 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} err="failed to get container status \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": rpc error: code = NotFound desc = could not find container \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": container with ID starting with de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.447301 4767 scope.go:117] "RemoveContainer" containerID="e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.447698 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} err="failed to get container status \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": rpc error: code = NotFound desc = could not find container \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": container with ID starting with e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.447724 4767 scope.go:117] "RemoveContainer" containerID="a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.448001 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} err="failed to get container status \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": rpc error: code = NotFound desc = could not find container \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": container with ID starting with a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.448099 4767 scope.go:117] "RemoveContainer" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.448473 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} err="failed to get container status \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": rpc error: code = NotFound desc = could not find container \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": container with ID starting with d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.448519 4767 scope.go:117] "RemoveContainer" containerID="1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.448750 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6"} err="failed to get container status \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": rpc error: code = NotFound desc = could not find container \"1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6\": container with ID starting with 1f22cb943ffe475628928289149d3f9f3da3c8a7d2c65837bbac44e0c616c0f6 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.448820 4767 scope.go:117] "RemoveContainer" containerID="f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449109 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5"} err="failed to get container status \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": rpc error: code = NotFound desc = could not find container \"f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5\": container with ID starting with f757ada7af29e6ba55e896a6775b371199ae636eff3d03bafadc638ab46949a5 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449156 4767 scope.go:117] "RemoveContainer" containerID="fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449392 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80"} err="failed to get container status \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": rpc error: code = NotFound desc = could not find container \"fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80\": container with ID starting with fd2f22cbc73df31c13ab10db268b3bb0d9d3c19161434fb4908c56c0a5a34e80 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449422 4767 scope.go:117] "RemoveContainer" containerID="0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449614 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e"} err="failed to get container status \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": rpc error: code = NotFound desc = could not find container \"0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e\": container with ID starting with 0c0b2760e7dba9e49b2e866aa215fcb3a7f38cf387e46d56db0d6bec4d60905e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449638 4767 scope.go:117] "RemoveContainer" containerID="fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449798 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570"} err="failed to get container status \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": rpc error: code = NotFound desc = could not find container \"fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570\": container with ID starting with fa2a5cfa49a0ef2e963687cb985f3972941a0d26b5512a592d0db7e93f3b2570 not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.449821 4767 scope.go:117] "RemoveContainer" containerID="8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.450146 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba"} err="failed to get container status \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": rpc error: code = NotFound desc = could not find container \"8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba\": container with ID starting with 8c67dab6c79ab0ef8dd88d879043431def54423147964acdad91ab78bd20cfba not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.450192 4767 scope.go:117] "RemoveContainer" containerID="de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.450425 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c"} err="failed to get container status \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": rpc error: code = NotFound desc = could not find container \"de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c\": container with ID starting with de1d562d48642b2722e048bd53fc57ba696ca60b58127dc3780db2a83261f67c not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.450452 4767 scope.go:117] "RemoveContainer" containerID="e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.450686 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e"} err="failed to get container status \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": rpc error: code = NotFound desc = could not find container \"e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e\": container with ID starting with e13c7ae24c465172b6a41d629b5cfcd627eef482a9d3d70d175d035006395e1e not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.450731 4767 scope.go:117] "RemoveContainer" containerID="a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.450999 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad"} err="failed to get container status \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": rpc error: code = NotFound desc = could not find container \"a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad\": container with ID starting with a238da655aacf317747f23f23d7c9656082bb83eecf592de63fbe2c0196380ad not found: ID does not exist" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.451051 4767 scope.go:117] "RemoveContainer" containerID="d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154" Jan 29 15:20:18 crc kubenswrapper[4767]: I0129 15:20:18.451294 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154"} err="failed to get container status \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": rpc error: code = NotFound desc = could not find container \"d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154\": container with ID starting with d76ca0318e04d79cf6c7e507750f02cfaa4f877f8d19fd50ebf18bc98d195154 not found: ID does not exist" Jan 29 15:20:19 crc kubenswrapper[4767]: I0129 15:20:19.025071 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="633e55cd-6740-4d13-84ef-023b691c0428" path="/var/lib/kubelet/pods/633e55cd-6740-4d13-84ef-023b691c0428/volumes" Jan 29 15:20:19 crc kubenswrapper[4767]: I0129 15:20:19.233370 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/2.log" Jan 29 15:20:19 crc kubenswrapper[4767]: I0129 15:20:19.233787 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/1.log" Jan 29 15:20:19 crc kubenswrapper[4767]: I0129 15:20:19.233874 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-9xcr9" event={"ID":"4615231a-e157-430e-9ecc-97c3fd7701bb","Type":"ContainerStarted","Data":"e36cd93904303db4a826d6066c616a4dd6c2ced445de38a81a3497171e437a71"} Jan 29 15:20:19 crc kubenswrapper[4767]: I0129 15:20:19.235124 4767 generic.go:334] "Generic (PLEG): container finished" podID="faaa8fe4-f480-4031-b6b7-c883d62209ee" containerID="bea5d79b9087a6267e6e69bdc822b3288105b63e8341b68aab1c3b98a49e8db8" exitCode=0 Jan 29 15:20:19 crc kubenswrapper[4767]: I0129 15:20:19.235161 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerDied","Data":"bea5d79b9087a6267e6e69bdc822b3288105b63e8341b68aab1c3b98a49e8db8"} Jan 29 15:20:19 crc kubenswrapper[4767]: I0129 15:20:19.235189 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"9fb4d6fb1f5f889a56f5c6cb9ff005a615808baf20108ee86eafbd47880fc01c"} Jan 29 15:20:20 crc kubenswrapper[4767]: I0129 15:20:20.243955 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"d31b1f1dee2ea0dc74a5ea5bf29f87d8a0312eacbf6655d2956dda12c3abe35e"} Jan 29 15:20:20 crc kubenswrapper[4767]: I0129 15:20:20.244623 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"c3c3be3d1a74290bc40f203128539d60517843630ba49ad511665d23e361cf93"} Jan 29 15:20:20 crc kubenswrapper[4767]: I0129 15:20:20.244638 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"cc8e6333f8b26ff9a09b74fd77db4b4be224049ee60cc53e413216c9e79d06b5"} Jan 29 15:20:20 crc kubenswrapper[4767]: I0129 15:20:20.244648 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"4c116f3740b64bbd5e591ca69be84c595e9770cbe7b4c8e2c62b04eb429c710e"} Jan 29 15:20:20 crc kubenswrapper[4767]: I0129 15:20:20.244656 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"be9a7ab6d78a8314bdd6015642ebb0b7b0eb491bcab72db9b75207283308c3da"} Jan 29 15:20:20 crc kubenswrapper[4767]: I0129 15:20:20.244664 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"b978a9c9775fe593595d308876a6e5b98f9f1f4b2267952c4504995e31fed148"} Jan 29 15:20:21 crc kubenswrapper[4767]: E0129 15:20:21.021737 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:20:22 crc kubenswrapper[4767]: I0129 15:20:22.258238 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"95dbb7cd7234c5261217c448cb5244c8b5a66076d2e5dcf32c41875409e254a8"} Jan 29 15:20:24 crc kubenswrapper[4767]: E0129 15:20:24.021737 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:20:25 crc kubenswrapper[4767]: I0129 15:20:25.279107 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" event={"ID":"faaa8fe4-f480-4031-b6b7-c883d62209ee","Type":"ContainerStarted","Data":"984669fde8c6e26d832807e8e249d77c87f987826993a96bf956585a004d7e4a"} Jan 29 15:20:25 crc kubenswrapper[4767]: I0129 15:20:25.280005 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:25 crc kubenswrapper[4767]: I0129 15:20:25.325711 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" podStartSLOduration=8.325684714 podStartE2EDuration="8.325684714s" podCreationTimestamp="2026-01-29 15:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:20:25.321293257 +0000 UTC m=+1181.131610326" watchObservedRunningTime="2026-01-29 15:20:25.325684714 +0000 UTC m=+1181.136001773" Jan 29 15:20:25 crc kubenswrapper[4767]: I0129 15:20:25.338835 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:26 crc kubenswrapper[4767]: I0129 15:20:26.284543 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:26 crc kubenswrapper[4767]: I0129 15:20:26.284675 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:26 crc kubenswrapper[4767]: I0129 15:20:26.321883 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:30 crc kubenswrapper[4767]: I0129 15:20:30.981595 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-qc6pz"] Jan 29 15:20:30 crc kubenswrapper[4767]: I0129 15:20:30.982854 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:30 crc kubenswrapper[4767]: I0129 15:20:30.985956 4767 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-pfvpk" Jan 29 15:20:30 crc kubenswrapper[4767]: I0129 15:20:30.986176 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Jan 29 15:20:30 crc kubenswrapper[4767]: I0129 15:20:30.986167 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Jan 29 15:20:30 crc kubenswrapper[4767]: I0129 15:20:30.986183 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.004697 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-qc6pz"] Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.114711 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1706f653-d300-4bb1-8e69-9f7de78ecb43-node-mnt\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.115051 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxffl\" (UniqueName: \"kubernetes.io/projected/1706f653-d300-4bb1-8e69-9f7de78ecb43-kube-api-access-lxffl\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.115194 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1706f653-d300-4bb1-8e69-9f7de78ecb43-crc-storage\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.218444 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1706f653-d300-4bb1-8e69-9f7de78ecb43-node-mnt\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.218572 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxffl\" (UniqueName: \"kubernetes.io/projected/1706f653-d300-4bb1-8e69-9f7de78ecb43-kube-api-access-lxffl\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.218726 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1706f653-d300-4bb1-8e69-9f7de78ecb43-crc-storage\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.219518 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1706f653-d300-4bb1-8e69-9f7de78ecb43-node-mnt\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.220487 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1706f653-d300-4bb1-8e69-9f7de78ecb43-crc-storage\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.249624 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxffl\" (UniqueName: \"kubernetes.io/projected/1706f653-d300-4bb1-8e69-9f7de78ecb43-kube-api-access-lxffl\") pod \"crc-storage-crc-qc6pz\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.327092 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:31 crc kubenswrapper[4767]: I0129 15:20:31.504007 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-qc6pz"] Jan 29 15:20:32 crc kubenswrapper[4767]: I0129 15:20:32.320429 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-qc6pz" event={"ID":"1706f653-d300-4bb1-8e69-9f7de78ecb43","Type":"ContainerStarted","Data":"31c3c1774de484c306fec933132bdb9277dd02230e1e22e5592355bd3f6e9b7e"} Jan 29 15:20:33 crc kubenswrapper[4767]: I0129 15:20:33.328263 4767 generic.go:334] "Generic (PLEG): container finished" podID="1706f653-d300-4bb1-8e69-9f7de78ecb43" containerID="16fa8e21cc5d3f0e1914cce83278281d6e6adbbca08624ae8279a85c41d4193e" exitCode=0 Jan 29 15:20:33 crc kubenswrapper[4767]: I0129 15:20:33.328328 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-qc6pz" event={"ID":"1706f653-d300-4bb1-8e69-9f7de78ecb43","Type":"ContainerDied","Data":"16fa8e21cc5d3f0e1914cce83278281d6e6adbbca08624ae8279a85c41d4193e"} Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.596252 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.764928 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1706f653-d300-4bb1-8e69-9f7de78ecb43-crc-storage\") pod \"1706f653-d300-4bb1-8e69-9f7de78ecb43\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.765024 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxffl\" (UniqueName: \"kubernetes.io/projected/1706f653-d300-4bb1-8e69-9f7de78ecb43-kube-api-access-lxffl\") pod \"1706f653-d300-4bb1-8e69-9f7de78ecb43\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.765052 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1706f653-d300-4bb1-8e69-9f7de78ecb43-node-mnt\") pod \"1706f653-d300-4bb1-8e69-9f7de78ecb43\" (UID: \"1706f653-d300-4bb1-8e69-9f7de78ecb43\") " Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.765416 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1706f653-d300-4bb1-8e69-9f7de78ecb43-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "1706f653-d300-4bb1-8e69-9f7de78ecb43" (UID: "1706f653-d300-4bb1-8e69-9f7de78ecb43"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.771127 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1706f653-d300-4bb1-8e69-9f7de78ecb43-kube-api-access-lxffl" (OuterVolumeSpecName: "kube-api-access-lxffl") pod "1706f653-d300-4bb1-8e69-9f7de78ecb43" (UID: "1706f653-d300-4bb1-8e69-9f7de78ecb43"). InnerVolumeSpecName "kube-api-access-lxffl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.786179 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1706f653-d300-4bb1-8e69-9f7de78ecb43-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "1706f653-d300-4bb1-8e69-9f7de78ecb43" (UID: "1706f653-d300-4bb1-8e69-9f7de78ecb43"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.867012 4767 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/1706f653-d300-4bb1-8e69-9f7de78ecb43-crc-storage\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.867078 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxffl\" (UniqueName: \"kubernetes.io/projected/1706f653-d300-4bb1-8e69-9f7de78ecb43-kube-api-access-lxffl\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:34 crc kubenswrapper[4767]: I0129 15:20:34.867101 4767 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/1706f653-d300-4bb1-8e69-9f7de78ecb43-node-mnt\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:35 crc kubenswrapper[4767]: E0129 15:20:35.025323 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-bv7wp" podUID="371621b9-a052-42ad-ba71-a37a3a064370" Jan 29 15:20:35 crc kubenswrapper[4767]: E0129 15:20:35.025754 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-86kzs" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" Jan 29 15:20:35 crc kubenswrapper[4767]: I0129 15:20:35.344996 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-qc6pz" event={"ID":"1706f653-d300-4bb1-8e69-9f7de78ecb43","Type":"ContainerDied","Data":"31c3c1774de484c306fec933132bdb9277dd02230e1e22e5592355bd3f6e9b7e"} Jan 29 15:20:35 crc kubenswrapper[4767]: I0129 15:20:35.345036 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31c3c1774de484c306fec933132bdb9277dd02230e1e22e5592355bd3f6e9b7e" Jan 29 15:20:35 crc kubenswrapper[4767]: I0129 15:20:35.345195 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-qc6pz" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.010172 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bv7wp"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.026443 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glb2r"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.026844 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="registry-server" containerID="cri-o://e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46" gracePeriod=30 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.033058 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86kzs"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.043045 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp7wv"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.043844 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tp7wv" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="registry-server" containerID="cri-o://2cb2ec7b012335fd18b42138e7d28da61f893b4ec981f29e4acb5ed4d36dfa42" gracePeriod=30 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.045937 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ftjvt"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.046185 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" podUID="0714bcea-20bf-4faf-b4b1-84df642268ac" containerName="marketplace-operator" containerID="cri-o://0b0b514a641d7663293829194c411d5f37d5f592b70affcb95faaeb03829e274" gracePeriod=30 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.059652 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2975f"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.059968 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="registry-server" containerID="cri-o://f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d" gracePeriod=30 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.068478 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-24ntq"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.068795 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-24ntq" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="registry-server" containerID="cri-o://5425a81c114abd07b63c498185cea3583adbde68d33a2a7d0cafb65b8459b0ae" gracePeriod=30 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.076747 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g7t6n"] Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.078431 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1706f653-d300-4bb1-8e69-9f7de78ecb43" containerName="storage" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.078457 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1706f653-d300-4bb1-8e69-9f7de78ecb43" containerName="storage" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.078612 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1706f653-d300-4bb1-8e69-9f7de78ecb43" containerName="storage" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.079087 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.085019 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g7t6n"] Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.252030 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc351b10-105e-46dd-bef5-b8a19abaad73-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.252089 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc351b10-105e-46dd-bef5-b8a19abaad73-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.252116 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6htbp\" (UniqueName: \"kubernetes.io/projected/dc351b10-105e-46dd-bef5-b8a19abaad73-kube-api-access-6htbp\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.353193 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc351b10-105e-46dd-bef5-b8a19abaad73-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.353236 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6htbp\" (UniqueName: \"kubernetes.io/projected/dc351b10-105e-46dd-bef5-b8a19abaad73-kube-api-access-6htbp\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.353315 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc351b10-105e-46dd-bef5-b8a19abaad73-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.354462 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc351b10-105e-46dd-bef5-b8a19abaad73-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.362742 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dc351b10-105e-46dd-bef5-b8a19abaad73-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.370419 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6htbp\" (UniqueName: \"kubernetes.io/projected/dc351b10-105e-46dd-bef5-b8a19abaad73-kube-api-access-6htbp\") pod \"marketplace-operator-79b997595-g7t6n\" (UID: \"dc351b10-105e-46dd-bef5-b8a19abaad73\") " pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.385271 4767 generic.go:334] "Generic (PLEG): container finished" podID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerID="5425a81c114abd07b63c498185cea3583adbde68d33a2a7d0cafb65b8459b0ae" exitCode=0 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.385350 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerDied","Data":"5425a81c114abd07b63c498185cea3583adbde68d33a2a7d0cafb65b8459b0ae"} Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.387485 4767 generic.go:334] "Generic (PLEG): container finished" podID="269bdfc1-3775-468a-a982-eebda8f7346e" containerID="f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d" exitCode=0 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.387532 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2975f" event={"ID":"269bdfc1-3775-468a-a982-eebda8f7346e","Type":"ContainerDied","Data":"f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d"} Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.389845 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-86kzs" event={"ID":"ac128fbf-9b98-49fe-9ab1-a20490c2529d","Type":"ContainerDied","Data":"3b926d27e7b605605e16394c01d5c98b3e783303f0da7c29758fa4a41172a60b"} Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.389936 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b926d27e7b605605e16394c01d5c98b3e783303f0da7c29758fa4a41172a60b" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.391520 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bv7wp" event={"ID":"371621b9-a052-42ad-ba71-a37a3a064370","Type":"ContainerDied","Data":"f5f660516a00e9a765e881551b69198b3407d0d7bc4602f7b7b963a4b4ef1fc4"} Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.391553 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5f660516a00e9a765e881551b69198b3407d0d7bc4602f7b7b963a4b4ef1fc4" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.393496 4767 generic.go:334] "Generic (PLEG): container finished" podID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerID="e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46" exitCode=0 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.393570 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glb2r" event={"ID":"0039b26d-8fc9-483e-9655-8ce3d9048727","Type":"ContainerDied","Data":"e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46"} Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.396217 4767 generic.go:334] "Generic (PLEG): container finished" podID="87659dc2-6bea-4599-8131-eaa87aee2406" containerID="2cb2ec7b012335fd18b42138e7d28da61f893b4ec981f29e4acb5ed4d36dfa42" exitCode=0 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.396263 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp7wv" event={"ID":"87659dc2-6bea-4599-8131-eaa87aee2406","Type":"ContainerDied","Data":"2cb2ec7b012335fd18b42138e7d28da61f893b4ec981f29e4acb5ed4d36dfa42"} Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.397788 4767 generic.go:334] "Generic (PLEG): container finished" podID="0714bcea-20bf-4faf-b4b1-84df642268ac" containerID="0b0b514a641d7663293829194c411d5f37d5f592b70affcb95faaeb03829e274" exitCode=0 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.397815 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" event={"ID":"0714bcea-20bf-4faf-b4b1-84df642268ac","Type":"ContainerDied","Data":"0b0b514a641d7663293829194c411d5f37d5f592b70affcb95faaeb03829e274"} Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.409392 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.488432 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d is running failed: container process not found" containerID="f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.489047 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d is running failed: container process not found" containerID="f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.489312 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d is running failed: container process not found" containerID="f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.489344 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-2975f" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="registry-server" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.596975 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-g7t6n"] Jan 29 15:20:40 crc kubenswrapper[4767]: W0129 15:20:40.653234 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc351b10_105e_46dd_bef5_b8a19abaad73.slice/crio-6244895ea010d9e105bbe598605e4668fd2a8b106f6bbac16e90a061d63dc2f4 WatchSource:0}: Error finding container 6244895ea010d9e105bbe598605e4668fd2a8b106f6bbac16e90a061d63dc2f4: Status 404 returned error can't find the container with id 6244895ea010d9e105bbe598605e4668fd2a8b106f6bbac16e90a061d63dc2f4 Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.658515 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.664354 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.679587 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46 is running failed: container process not found" containerID="e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.680695 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46 is running failed: container process not found" containerID="e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.681191 4767 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46 is running failed: container process not found" containerID="e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46" cmd=["grpc_health_probe","-addr=:50051"] Jan 29 15:20:40 crc kubenswrapper[4767]: E0129 15:20:40.681231 4767 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-glb2r" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="registry-server" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.757585 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b68h\" (UniqueName: \"kubernetes.io/projected/371621b9-a052-42ad-ba71-a37a3a064370-kube-api-access-5b68h\") pod \"371621b9-a052-42ad-ba71-a37a3a064370\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.757632 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-utilities\") pod \"371621b9-a052-42ad-ba71-a37a3a064370\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.757717 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-catalog-content\") pod \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.757754 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swgsv\" (UniqueName: \"kubernetes.io/projected/ac128fbf-9b98-49fe-9ab1-a20490c2529d-kube-api-access-swgsv\") pod \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.757783 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-utilities\") pod \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\" (UID: \"ac128fbf-9b98-49fe-9ab1-a20490c2529d\") " Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.757797 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-catalog-content\") pod \"371621b9-a052-42ad-ba71-a37a3a064370\" (UID: \"371621b9-a052-42ad-ba71-a37a3a064370\") " Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.758266 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac128fbf-9b98-49fe-9ab1-a20490c2529d" (UID: "ac128fbf-9b98-49fe-9ab1-a20490c2529d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.759029 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "371621b9-a052-42ad-ba71-a37a3a064370" (UID: "371621b9-a052-42ad-ba71-a37a3a064370"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.759436 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-utilities" (OuterVolumeSpecName: "utilities") pod "ac128fbf-9b98-49fe-9ab1-a20490c2529d" (UID: "ac128fbf-9b98-49fe-9ab1-a20490c2529d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.759803 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-utilities" (OuterVolumeSpecName: "utilities") pod "371621b9-a052-42ad-ba71-a37a3a064370" (UID: "371621b9-a052-42ad-ba71-a37a3a064370"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.769074 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac128fbf-9b98-49fe-9ab1-a20490c2529d-kube-api-access-swgsv" (OuterVolumeSpecName: "kube-api-access-swgsv") pod "ac128fbf-9b98-49fe-9ab1-a20490c2529d" (UID: "ac128fbf-9b98-49fe-9ab1-a20490c2529d"). InnerVolumeSpecName "kube-api-access-swgsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.769461 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/371621b9-a052-42ad-ba71-a37a3a064370-kube-api-access-5b68h" (OuterVolumeSpecName: "kube-api-access-5b68h") pod "371621b9-a052-42ad-ba71-a37a3a064370" (UID: "371621b9-a052-42ad-ba71-a37a3a064370"). InnerVolumeSpecName "kube-api-access-5b68h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.859079 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.859482 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swgsv\" (UniqueName: \"kubernetes.io/projected/ac128fbf-9b98-49fe-9ab1-a20490c2529d-kube-api-access-swgsv\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.859499 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac128fbf-9b98-49fe-9ab1-a20490c2529d-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.859701 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.859715 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b68h\" (UniqueName: \"kubernetes.io/projected/371621b9-a052-42ad-ba71-a37a3a064370-kube-api-access-5b68h\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.859726 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/371621b9-a052-42ad-ba71-a37a3a064370-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.894133 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.934751 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.941227 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:20:40 crc kubenswrapper[4767]: I0129 15:20:40.944786 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.013715 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060771 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-utilities\") pod \"8579daf0-a4b5-473c-9133-048eaa57c5bb\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060832 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62mwb\" (UniqueName: \"kubernetes.io/projected/8579daf0-a4b5-473c-9133-048eaa57c5bb-kube-api-access-62mwb\") pod \"8579daf0-a4b5-473c-9133-048eaa57c5bb\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060854 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-trusted-ca\") pod \"0714bcea-20bf-4faf-b4b1-84df642268ac\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060870 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-utilities\") pod \"87659dc2-6bea-4599-8131-eaa87aee2406\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060903 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-catalog-content\") pod \"87659dc2-6bea-4599-8131-eaa87aee2406\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060936 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-operator-metrics\") pod \"0714bcea-20bf-4faf-b4b1-84df642268ac\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060962 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-catalog-content\") pod \"8579daf0-a4b5-473c-9133-048eaa57c5bb\" (UID: \"8579daf0-a4b5-473c-9133-048eaa57c5bb\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.060987 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffkxn\" (UniqueName: \"kubernetes.io/projected/87659dc2-6bea-4599-8131-eaa87aee2406-kube-api-access-ffkxn\") pod \"87659dc2-6bea-4599-8131-eaa87aee2406\" (UID: \"87659dc2-6bea-4599-8131-eaa87aee2406\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.061013 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx79c\" (UniqueName: \"kubernetes.io/projected/0714bcea-20bf-4faf-b4b1-84df642268ac-kube-api-access-hx79c\") pod \"0714bcea-20bf-4faf-b4b1-84df642268ac\" (UID: \"0714bcea-20bf-4faf-b4b1-84df642268ac\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.061038 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdxfw\" (UniqueName: \"kubernetes.io/projected/0039b26d-8fc9-483e-9655-8ce3d9048727-kube-api-access-tdxfw\") pod \"0039b26d-8fc9-483e-9655-8ce3d9048727\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.061067 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-utilities\") pod \"0039b26d-8fc9-483e-9655-8ce3d9048727\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.061151 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-catalog-content\") pod \"0039b26d-8fc9-483e-9655-8ce3d9048727\" (UID: \"0039b26d-8fc9-483e-9655-8ce3d9048727\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.061870 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-utilities" (OuterVolumeSpecName: "utilities") pod "87659dc2-6bea-4599-8131-eaa87aee2406" (UID: "87659dc2-6bea-4599-8131-eaa87aee2406"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.062392 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0714bcea-20bf-4faf-b4b1-84df642268ac" (UID: "0714bcea-20bf-4faf-b4b1-84df642268ac"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.062509 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-utilities" (OuterVolumeSpecName: "utilities") pod "8579daf0-a4b5-473c-9133-048eaa57c5bb" (UID: "8579daf0-a4b5-473c-9133-048eaa57c5bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.063420 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-utilities" (OuterVolumeSpecName: "utilities") pod "0039b26d-8fc9-483e-9655-8ce3d9048727" (UID: "0039b26d-8fc9-483e-9655-8ce3d9048727"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.067443 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0039b26d-8fc9-483e-9655-8ce3d9048727-kube-api-access-tdxfw" (OuterVolumeSpecName: "kube-api-access-tdxfw") pod "0039b26d-8fc9-483e-9655-8ce3d9048727" (UID: "0039b26d-8fc9-483e-9655-8ce3d9048727"). InnerVolumeSpecName "kube-api-access-tdxfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.067488 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87659dc2-6bea-4599-8131-eaa87aee2406-kube-api-access-ffkxn" (OuterVolumeSpecName: "kube-api-access-ffkxn") pod "87659dc2-6bea-4599-8131-eaa87aee2406" (UID: "87659dc2-6bea-4599-8131-eaa87aee2406"). InnerVolumeSpecName "kube-api-access-ffkxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.067587 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0714bcea-20bf-4faf-b4b1-84df642268ac-kube-api-access-hx79c" (OuterVolumeSpecName: "kube-api-access-hx79c") pod "0714bcea-20bf-4faf-b4b1-84df642268ac" (UID: "0714bcea-20bf-4faf-b4b1-84df642268ac"). InnerVolumeSpecName "kube-api-access-hx79c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.067669 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8579daf0-a4b5-473c-9133-048eaa57c5bb-kube-api-access-62mwb" (OuterVolumeSpecName: "kube-api-access-62mwb") pod "8579daf0-a4b5-473c-9133-048eaa57c5bb" (UID: "8579daf0-a4b5-473c-9133-048eaa57c5bb"). InnerVolumeSpecName "kube-api-access-62mwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.094072 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0714bcea-20bf-4faf-b4b1-84df642268ac" (UID: "0714bcea-20bf-4faf-b4b1-84df642268ac"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.129856 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0039b26d-8fc9-483e-9655-8ce3d9048727" (UID: "0039b26d-8fc9-483e-9655-8ce3d9048727"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.132493 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87659dc2-6bea-4599-8131-eaa87aee2406" (UID: "87659dc2-6bea-4599-8131-eaa87aee2406"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167409 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v25cq\" (UniqueName: \"kubernetes.io/projected/269bdfc1-3775-468a-a982-eebda8f7346e-kube-api-access-v25cq\") pod \"269bdfc1-3775-468a-a982-eebda8f7346e\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167484 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-utilities\") pod \"269bdfc1-3775-468a-a982-eebda8f7346e\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167555 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-catalog-content\") pod \"269bdfc1-3775-468a-a982-eebda8f7346e\" (UID: \"269bdfc1-3775-468a-a982-eebda8f7346e\") " Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167798 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62mwb\" (UniqueName: \"kubernetes.io/projected/8579daf0-a4b5-473c-9133-048eaa57c5bb-kube-api-access-62mwb\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167818 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167827 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167835 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87659dc2-6bea-4599-8131-eaa87aee2406-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167844 4767 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0714bcea-20bf-4faf-b4b1-84df642268ac-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167853 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffkxn\" (UniqueName: \"kubernetes.io/projected/87659dc2-6bea-4599-8131-eaa87aee2406-kube-api-access-ffkxn\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167862 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx79c\" (UniqueName: \"kubernetes.io/projected/0714bcea-20bf-4faf-b4b1-84df642268ac-kube-api-access-hx79c\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167870 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdxfw\" (UniqueName: \"kubernetes.io/projected/0039b26d-8fc9-483e-9655-8ce3d9048727-kube-api-access-tdxfw\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167878 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167886 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0039b26d-8fc9-483e-9655-8ce3d9048727-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.167911 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.169367 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-utilities" (OuterVolumeSpecName: "utilities") pod "269bdfc1-3775-468a-a982-eebda8f7346e" (UID: "269bdfc1-3775-468a-a982-eebda8f7346e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.174531 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/269bdfc1-3775-468a-a982-eebda8f7346e-kube-api-access-v25cq" (OuterVolumeSpecName: "kube-api-access-v25cq") pod "269bdfc1-3775-468a-a982-eebda8f7346e" (UID: "269bdfc1-3775-468a-a982-eebda8f7346e"). InnerVolumeSpecName "kube-api-access-v25cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.194094 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "269bdfc1-3775-468a-a982-eebda8f7346e" (UID: "269bdfc1-3775-468a-a982-eebda8f7346e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.213824 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8579daf0-a4b5-473c-9133-048eaa57c5bb" (UID: "8579daf0-a4b5-473c-9133-048eaa57c5bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.268876 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v25cq\" (UniqueName: \"kubernetes.io/projected/269bdfc1-3775-468a-a982-eebda8f7346e-kube-api-access-v25cq\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.268936 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.268949 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8579daf0-a4b5-473c-9133-048eaa57c5bb-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.268958 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269bdfc1-3775-468a-a982-eebda8f7346e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.405426 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" event={"ID":"0714bcea-20bf-4faf-b4b1-84df642268ac","Type":"ContainerDied","Data":"f53fa6aa6b82d2959119f220c20968a71054c7f12e182f6eb0352bf19df451ae"} Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.405509 4767 scope.go:117] "RemoveContainer" containerID="0b0b514a641d7663293829194c411d5f37d5f592b70affcb95faaeb03829e274" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.405574 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ftjvt" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.408203 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-24ntq" event={"ID":"8579daf0-a4b5-473c-9133-048eaa57c5bb","Type":"ContainerDied","Data":"d9663fe41250c3c6c980ea39b239067dcb1596918b5d6f1517b555675ce30a3d"} Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.408237 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-24ntq" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.412815 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2975f" event={"ID":"269bdfc1-3775-468a-a982-eebda8f7346e","Type":"ContainerDied","Data":"bd8f0028e3046fdb498c3509a9644df610648749dd5640929da6fec8b36cd823"} Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.413116 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2975f" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.415743 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" event={"ID":"dc351b10-105e-46dd-bef5-b8a19abaad73","Type":"ContainerStarted","Data":"358c7c20bb768370fbbbd4b4ed0065a9dc1299064f027ad38b87c5f1dc2e631b"} Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.415796 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" event={"ID":"dc351b10-105e-46dd-bef5-b8a19abaad73","Type":"ContainerStarted","Data":"6244895ea010d9e105bbe598605e4668fd2a8b106f6bbac16e90a061d63dc2f4"} Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.416071 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.424891 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.425557 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-glb2r" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.425564 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-glb2r" event={"ID":"0039b26d-8fc9-483e-9655-8ce3d9048727","Type":"ContainerDied","Data":"e8bf714f123467db3df8c5397daf88077fd8bcfad5fd80cc02f35229aa4f5f78"} Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.429165 4767 scope.go:117] "RemoveContainer" containerID="5425a81c114abd07b63c498185cea3583adbde68d33a2a7d0cafb65b8459b0ae" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.429664 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-86kzs" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.431615 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tp7wv" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.434205 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bv7wp" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.434403 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tp7wv" event={"ID":"87659dc2-6bea-4599-8131-eaa87aee2406","Type":"ContainerDied","Data":"1bf083074d437ac09c0afe62491036cc5911996756aaa65b39db68f76fd58150"} Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.446818 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-g7t6n" podStartSLOduration=1.446793946 podStartE2EDuration="1.446793946s" podCreationTimestamp="2026-01-29 15:20:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:20:41.440418146 +0000 UTC m=+1197.250735185" watchObservedRunningTime="2026-01-29 15:20:41.446793946 +0000 UTC m=+1197.257111005" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.455135 4767 scope.go:117] "RemoveContainer" containerID="837e6fa44a173a70075a6b7c45bbaa0f698e024f965a5f2bf5d245b37ee6851d" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.499808 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-glb2r"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.503073 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-glb2r"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.528266 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2975f"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.532623 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2975f"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.546585 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ftjvt"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.546969 4767 scope.go:117] "RemoveContainer" containerID="1cc1111c5e3e0566a4412462e20991b0e247b95575f38cb22ba51af108a83ce5" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.548301 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ftjvt"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.556363 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tp7wv"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.564484 4767 scope.go:117] "RemoveContainer" containerID="f96776603be95202a8bd638caa9dff04024905cf01a40262f5949e957815b44d" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.565534 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tp7wv"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.590203 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bv7wp"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.591808 4767 scope.go:117] "RemoveContainer" containerID="8401f3dc733c542521518b734aaaf0b514baf17d2d9948b954ca6c68e13f7431" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.600924 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bv7wp"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.615105 4767 scope.go:117] "RemoveContainer" containerID="2a37e58d2297c3d114f228f518857857c919fc3b13486f89629b238aae8d3f3f" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.620834 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-24ntq"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.629174 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-24ntq"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.643311 4767 scope.go:117] "RemoveContainer" containerID="e619ec93b34e3f814563aa582a0de8ff6beb5ecf3bb934ba0cdfa92619611b46" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.650741 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-86kzs"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.656732 4767 scope.go:117] "RemoveContainer" containerID="0bb6129f1feb35a6265aa8cd32b154087d97aaaa46d047f1c65276bcd5be0271" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.657207 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-86kzs"] Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.671573 4767 scope.go:117] "RemoveContainer" containerID="3498a85f1679f1407544bfa63dfc8cffde177a7f016e3cee801d02b5a0e15cd1" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.684667 4767 scope.go:117] "RemoveContainer" containerID="2cb2ec7b012335fd18b42138e7d28da61f893b4ec981f29e4acb5ed4d36dfa42" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.699108 4767 scope.go:117] "RemoveContainer" containerID="53d1cb912b439b32e9b89ff254f6d2033ccf5272f466035d890cb1c0738c3f4b" Jan 29 15:20:41 crc kubenswrapper[4767]: I0129 15:20:41.714238 4767 scope.go:117] "RemoveContainer" containerID="d0fdf6efa5c6c883084f5ba572b581986b85c832f766c65f70b40715dabaa58c" Jan 29 15:20:42 crc kubenswrapper[4767]: I0129 15:20:42.805724 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:20:42 crc kubenswrapper[4767]: I0129 15:20:42.805789 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.025292 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" path="/var/lib/kubelet/pods/0039b26d-8fc9-483e-9655-8ce3d9048727/volumes" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.026584 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0714bcea-20bf-4faf-b4b1-84df642268ac" path="/var/lib/kubelet/pods/0714bcea-20bf-4faf-b4b1-84df642268ac/volumes" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.027096 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" path="/var/lib/kubelet/pods/269bdfc1-3775-468a-a982-eebda8f7346e/volumes" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.028331 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="371621b9-a052-42ad-ba71-a37a3a064370" path="/var/lib/kubelet/pods/371621b9-a052-42ad-ba71-a37a3a064370/volumes" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.028798 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" path="/var/lib/kubelet/pods/8579daf0-a4b5-473c-9133-048eaa57c5bb/volumes" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.029417 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" path="/var/lib/kubelet/pods/87659dc2-6bea-4599-8131-eaa87aee2406/volumes" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.030530 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" path="/var/lib/kubelet/pods/ac128fbf-9b98-49fe-9ab1-a20490c2529d/volumes" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.243696 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c9h8z"] Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.243984 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244004 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244028 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="371621b9-a052-42ad-ba71-a37a3a064370" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244039 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="371621b9-a052-42ad-ba71-a37a3a064370" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244055 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244066 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244082 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244090 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244116 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244126 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244142 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244152 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244167 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244179 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244195 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244204 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244217 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244225 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244235 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0714bcea-20bf-4faf-b4b1-84df642268ac" containerName="marketplace-operator" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244245 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0714bcea-20bf-4faf-b4b1-84df642268ac" containerName="marketplace-operator" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244260 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244268 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244279 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244287 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="extract-content" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244304 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244316 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244330 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244338 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: E0129 15:20:43.244352 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244361 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244495 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0714bcea-20bf-4faf-b4b1-84df642268ac" containerName="marketplace-operator" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244509 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="87659dc2-6bea-4599-8131-eaa87aee2406" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244524 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0039b26d-8fc9-483e-9655-8ce3d9048727" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244535 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="371621b9-a052-42ad-ba71-a37a3a064370" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244553 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8579daf0-a4b5-473c-9133-048eaa57c5bb" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244566 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="269bdfc1-3775-468a-a982-eebda8f7346e" containerName="registry-server" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.244577 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac128fbf-9b98-49fe-9ab1-a20490c2529d" containerName="extract-utilities" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.245466 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.246932 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9h8z"] Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.248712 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.303220 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdqvl\" (UniqueName: \"kubernetes.io/projected/3ff6370b-64ad-49cf-8f74-eb50bb30d826-kube-api-access-vdqvl\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.303298 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ff6370b-64ad-49cf-8f74-eb50bb30d826-utilities\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.303352 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ff6370b-64ad-49cf-8f74-eb50bb30d826-catalog-content\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.404727 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdqvl\" (UniqueName: \"kubernetes.io/projected/3ff6370b-64ad-49cf-8f74-eb50bb30d826-kube-api-access-vdqvl\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.404787 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ff6370b-64ad-49cf-8f74-eb50bb30d826-utilities\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.404830 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ff6370b-64ad-49cf-8f74-eb50bb30d826-catalog-content\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.405382 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ff6370b-64ad-49cf-8f74-eb50bb30d826-catalog-content\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.405786 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ff6370b-64ad-49cf-8f74-eb50bb30d826-utilities\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.438095 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdqvl\" (UniqueName: \"kubernetes.io/projected/3ff6370b-64ad-49cf-8f74-eb50bb30d826-kube-api-access-vdqvl\") pod \"redhat-marketplace-c9h8z\" (UID: \"3ff6370b-64ad-49cf-8f74-eb50bb30d826\") " pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.441446 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sp2xf"] Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.442812 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.445309 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.454126 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sp2xf"] Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.506104 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-utilities\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.506196 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-catalog-content\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.506243 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw74x\" (UniqueName: \"kubernetes.io/projected/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-kube-api-access-tw74x\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.570520 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.607341 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-utilities\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.607430 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-catalog-content\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.607539 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw74x\" (UniqueName: \"kubernetes.io/projected/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-kube-api-access-tw74x\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.608043 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-catalog-content\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.608130 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-utilities\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.630554 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw74x\" (UniqueName: \"kubernetes.io/projected/e635388d-4adc-4dfe-aca6-4ed3c7ecc646-kube-api-access-tw74x\") pod \"redhat-operators-sp2xf\" (UID: \"e635388d-4adc-4dfe-aca6-4ed3c7ecc646\") " pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.759957 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c9h8z"] Jan 29 15:20:43 crc kubenswrapper[4767]: W0129 15:20:43.768071 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ff6370b_64ad_49cf_8f74_eb50bb30d826.slice/crio-504d4997edc54dfc85f51a2d745b72cc3bb7cbc77e8acdd9ba1b0ede72fafeb2 WatchSource:0}: Error finding container 504d4997edc54dfc85f51a2d745b72cc3bb7cbc77e8acdd9ba1b0ede72fafeb2: Status 404 returned error can't find the container with id 504d4997edc54dfc85f51a2d745b72cc3bb7cbc77e8acdd9ba1b0ede72fafeb2 Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.775631 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:43 crc kubenswrapper[4767]: I0129 15:20:43.954955 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sp2xf"] Jan 29 15:20:43 crc kubenswrapper[4767]: W0129 15:20:43.987964 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode635388d_4adc_4dfe_aca6_4ed3c7ecc646.slice/crio-a87b797f931532730ec88ac94e6c9e6983bca35706e6ba4e8428e9b542504064 WatchSource:0}: Error finding container a87b797f931532730ec88ac94e6c9e6983bca35706e6ba4e8428e9b542504064: Status 404 returned error can't find the container with id a87b797f931532730ec88ac94e6c9e6983bca35706e6ba4e8428e9b542504064 Jan 29 15:20:44 crc kubenswrapper[4767]: I0129 15:20:44.463441 4767 generic.go:334] "Generic (PLEG): container finished" podID="3ff6370b-64ad-49cf-8f74-eb50bb30d826" containerID="14d865f435f2125763abe1623ec001d21cef9dfa18cf31ad9a5d8359eba50d21" exitCode=0 Jan 29 15:20:44 crc kubenswrapper[4767]: I0129 15:20:44.463527 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9h8z" event={"ID":"3ff6370b-64ad-49cf-8f74-eb50bb30d826","Type":"ContainerDied","Data":"14d865f435f2125763abe1623ec001d21cef9dfa18cf31ad9a5d8359eba50d21"} Jan 29 15:20:44 crc kubenswrapper[4767]: I0129 15:20:44.463570 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9h8z" event={"ID":"3ff6370b-64ad-49cf-8f74-eb50bb30d826","Type":"ContainerStarted","Data":"504d4997edc54dfc85f51a2d745b72cc3bb7cbc77e8acdd9ba1b0ede72fafeb2"} Jan 29 15:20:44 crc kubenswrapper[4767]: I0129 15:20:44.465934 4767 generic.go:334] "Generic (PLEG): container finished" podID="e635388d-4adc-4dfe-aca6-4ed3c7ecc646" containerID="7d5c55658269213f9b744641d452b5d70fdbb5d592ef531bbe9b348420571299" exitCode=0 Jan 29 15:20:44 crc kubenswrapper[4767]: I0129 15:20:44.466009 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sp2xf" event={"ID":"e635388d-4adc-4dfe-aca6-4ed3c7ecc646","Type":"ContainerDied","Data":"7d5c55658269213f9b744641d452b5d70fdbb5d592ef531bbe9b348420571299"} Jan 29 15:20:44 crc kubenswrapper[4767]: I0129 15:20:44.466039 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sp2xf" event={"ID":"e635388d-4adc-4dfe-aca6-4ed3c7ecc646","Type":"ContainerStarted","Data":"a87b797f931532730ec88ac94e6c9e6983bca35706e6ba4e8428e9b542504064"} Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.433636 4767 scope.go:117] "RemoveContainer" containerID="37f6007e8a02944007adfddef1dc2aa4d37cf94a699a05acaeff2bc86a4e657b" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.642524 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tst6c"] Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.644882 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.649564 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.650530 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tst6c"] Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.734196 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ef69eb-5928-4466-988d-6d4955721b2e-utilities\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.734273 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdmjl\" (UniqueName: \"kubernetes.io/projected/e1ef69eb-5928-4466-988d-6d4955721b2e-kube-api-access-jdmjl\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.734357 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ef69eb-5928-4466-988d-6d4955721b2e-catalog-content\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.829647 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zn5dm"] Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.830650 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.832582 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.835463 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ef69eb-5928-4466-988d-6d4955721b2e-utilities\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.835518 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdmjl\" (UniqueName: \"kubernetes.io/projected/e1ef69eb-5928-4466-988d-6d4955721b2e-kube-api-access-jdmjl\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.835588 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ef69eb-5928-4466-988d-6d4955721b2e-catalog-content\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.836104 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1ef69eb-5928-4466-988d-6d4955721b2e-catalog-content\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.836288 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1ef69eb-5928-4466-988d-6d4955721b2e-utilities\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.839527 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zn5dm"] Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.889608 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdmjl\" (UniqueName: \"kubernetes.io/projected/e1ef69eb-5928-4466-988d-6d4955721b2e-kube-api-access-jdmjl\") pod \"community-operators-tst6c\" (UID: \"e1ef69eb-5928-4466-988d-6d4955721b2e\") " pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.937479 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2zf7\" (UniqueName: \"kubernetes.io/projected/149720b9-74d1-432e-ad66-7824a18ff8a5-kube-api-access-w2zf7\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.937576 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149720b9-74d1-432e-ad66-7824a18ff8a5-catalog-content\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:45 crc kubenswrapper[4767]: I0129 15:20:45.937632 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149720b9-74d1-432e-ad66-7824a18ff8a5-utilities\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.020263 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.039102 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149720b9-74d1-432e-ad66-7824a18ff8a5-catalog-content\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.039211 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149720b9-74d1-432e-ad66-7824a18ff8a5-utilities\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.039310 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2zf7\" (UniqueName: \"kubernetes.io/projected/149720b9-74d1-432e-ad66-7824a18ff8a5-kube-api-access-w2zf7\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.040387 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149720b9-74d1-432e-ad66-7824a18ff8a5-utilities\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.040381 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149720b9-74d1-432e-ad66-7824a18ff8a5-catalog-content\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.072184 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2zf7\" (UniqueName: \"kubernetes.io/projected/149720b9-74d1-432e-ad66-7824a18ff8a5-kube-api-access-w2zf7\") pod \"certified-operators-zn5dm\" (UID: \"149720b9-74d1-432e-ad66-7824a18ff8a5\") " pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.191503 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.426463 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tst6c"] Jan 29 15:20:46 crc kubenswrapper[4767]: W0129 15:20:46.441764 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1ef69eb_5928_4466_988d_6d4955721b2e.slice/crio-59afbdaccbfd7ba35f020fa06b131c2d7b2bdb851235d746c9e9daef956331b6 WatchSource:0}: Error finding container 59afbdaccbfd7ba35f020fa06b131c2d7b2bdb851235d746c9e9daef956331b6: Status 404 returned error can't find the container with id 59afbdaccbfd7ba35f020fa06b131c2d7b2bdb851235d746c9e9daef956331b6 Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.480821 4767 generic.go:334] "Generic (PLEG): container finished" podID="e635388d-4adc-4dfe-aca6-4ed3c7ecc646" containerID="597a3910c0dfd11fda44664ad46cd92258295b93767f26d7ff4b6b28f33ac0f4" exitCode=0 Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.480920 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sp2xf" event={"ID":"e635388d-4adc-4dfe-aca6-4ed3c7ecc646","Type":"ContainerDied","Data":"597a3910c0dfd11fda44664ad46cd92258295b93767f26d7ff4b6b28f33ac0f4"} Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.485579 4767 generic.go:334] "Generic (PLEG): container finished" podID="3ff6370b-64ad-49cf-8f74-eb50bb30d826" containerID="d69cb63901256ebbb4422facf5152147dcefa3183f727cbafaca69f3224a113c" exitCode=0 Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.485652 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9h8z" event={"ID":"3ff6370b-64ad-49cf-8f74-eb50bb30d826","Type":"ContainerDied","Data":"d69cb63901256ebbb4422facf5152147dcefa3183f727cbafaca69f3224a113c"} Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.487642 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tst6c" event={"ID":"e1ef69eb-5928-4466-988d-6d4955721b2e","Type":"ContainerStarted","Data":"59afbdaccbfd7ba35f020fa06b131c2d7b2bdb851235d746c9e9daef956331b6"} Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.492682 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-9xcr9_4615231a-e157-430e-9ecc-97c3fd7701bb/kube-multus/2.log" Jan 29 15:20:46 crc kubenswrapper[4767]: I0129 15:20:46.572497 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zn5dm"] Jan 29 15:20:46 crc kubenswrapper[4767]: W0129 15:20:46.579693 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod149720b9_74d1_432e_ad66_7824a18ff8a5.slice/crio-ae94c65b2de4ab5f6a8179faa6e7480260378361f1e09e1c6d6df05ce66d82b6 WatchSource:0}: Error finding container ae94c65b2de4ab5f6a8179faa6e7480260378361f1e09e1c6d6df05ce66d82b6: Status 404 returned error can't find the container with id ae94c65b2de4ab5f6a8179faa6e7480260378361f1e09e1c6d6df05ce66d82b6 Jan 29 15:20:47 crc kubenswrapper[4767]: I0129 15:20:47.505397 4767 generic.go:334] "Generic (PLEG): container finished" podID="e1ef69eb-5928-4466-988d-6d4955721b2e" containerID="57488d6a89f760a0db73a8eb1afcba19421a0322cd0d6baeb4dd8da9aedc3c4b" exitCode=0 Jan 29 15:20:47 crc kubenswrapper[4767]: I0129 15:20:47.505772 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tst6c" event={"ID":"e1ef69eb-5928-4466-988d-6d4955721b2e","Type":"ContainerDied","Data":"57488d6a89f760a0db73a8eb1afcba19421a0322cd0d6baeb4dd8da9aedc3c4b"} Jan 29 15:20:47 crc kubenswrapper[4767]: I0129 15:20:47.508563 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:20:47 crc kubenswrapper[4767]: I0129 15:20:47.511663 4767 generic.go:334] "Generic (PLEG): container finished" podID="149720b9-74d1-432e-ad66-7824a18ff8a5" containerID="c9177120a19817520779986de4e6f6d5eb7036b05bda585d690a67729463fc72" exitCode=0 Jan 29 15:20:47 crc kubenswrapper[4767]: I0129 15:20:47.511696 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn5dm" event={"ID":"149720b9-74d1-432e-ad66-7824a18ff8a5","Type":"ContainerDied","Data":"c9177120a19817520779986de4e6f6d5eb7036b05bda585d690a67729463fc72"} Jan 29 15:20:47 crc kubenswrapper[4767]: I0129 15:20:47.511723 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn5dm" event={"ID":"149720b9-74d1-432e-ad66-7824a18ff8a5","Type":"ContainerStarted","Data":"ae94c65b2de4ab5f6a8179faa6e7480260378361f1e09e1c6d6df05ce66d82b6"} Jan 29 15:20:48 crc kubenswrapper[4767]: I0129 15:20:48.353813 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kqqcl" Jan 29 15:20:48 crc kubenswrapper[4767]: I0129 15:20:48.521735 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn5dm" event={"ID":"149720b9-74d1-432e-ad66-7824a18ff8a5","Type":"ContainerStarted","Data":"562499fe2d440398e47cfc47a5b25e015075a037fe844e769215d795dceffbda"} Jan 29 15:20:48 crc kubenswrapper[4767]: I0129 15:20:48.526222 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c9h8z" event={"ID":"3ff6370b-64ad-49cf-8f74-eb50bb30d826","Type":"ContainerStarted","Data":"8fb6ba0b2c9ec7fe680c984201ea2ea67c614e16cfc0c077f1d60e7c59d11772"} Jan 29 15:20:48 crc kubenswrapper[4767]: I0129 15:20:48.528814 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sp2xf" event={"ID":"e635388d-4adc-4dfe-aca6-4ed3c7ecc646","Type":"ContainerStarted","Data":"16c55b8a7a21713a43c5a84770dc855b0a1ffead612cde27d393408e7cdfd963"} Jan 29 15:20:48 crc kubenswrapper[4767]: I0129 15:20:48.530805 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tst6c" event={"ID":"e1ef69eb-5928-4466-988d-6d4955721b2e","Type":"ContainerStarted","Data":"0039ba3c794fb38108bbe1bfd7777e5cbc69e11c18f20191cd06bb85b36515c5"} Jan 29 15:20:48 crc kubenswrapper[4767]: I0129 15:20:48.561281 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c9h8z" podStartSLOduration=2.436482082 podStartE2EDuration="5.561255174s" podCreationTimestamp="2026-01-29 15:20:43 +0000 UTC" firstStartedPulling="2026-01-29 15:20:44.466340929 +0000 UTC m=+1200.276657958" lastFinishedPulling="2026-01-29 15:20:47.591114011 +0000 UTC m=+1203.401431050" observedRunningTime="2026-01-29 15:20:48.55800123 +0000 UTC m=+1204.368318269" watchObservedRunningTime="2026-01-29 15:20:48.561255174 +0000 UTC m=+1204.371572213" Jan 29 15:20:48 crc kubenswrapper[4767]: I0129 15:20:48.597699 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sp2xf" podStartSLOduration=2.51972075 podStartE2EDuration="5.597672219s" podCreationTimestamp="2026-01-29 15:20:43 +0000 UTC" firstStartedPulling="2026-01-29 15:20:44.467955385 +0000 UTC m=+1200.278272464" lastFinishedPulling="2026-01-29 15:20:47.545906894 +0000 UTC m=+1203.356223933" observedRunningTime="2026-01-29 15:20:48.593221891 +0000 UTC m=+1204.403538930" watchObservedRunningTime="2026-01-29 15:20:48.597672219 +0000 UTC m=+1204.407989268" Jan 29 15:20:49 crc kubenswrapper[4767]: I0129 15:20:49.539621 4767 generic.go:334] "Generic (PLEG): container finished" podID="149720b9-74d1-432e-ad66-7824a18ff8a5" containerID="562499fe2d440398e47cfc47a5b25e015075a037fe844e769215d795dceffbda" exitCode=0 Jan 29 15:20:49 crc kubenswrapper[4767]: I0129 15:20:49.539753 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn5dm" event={"ID":"149720b9-74d1-432e-ad66-7824a18ff8a5","Type":"ContainerDied","Data":"562499fe2d440398e47cfc47a5b25e015075a037fe844e769215d795dceffbda"} Jan 29 15:20:49 crc kubenswrapper[4767]: I0129 15:20:49.543075 4767 generic.go:334] "Generic (PLEG): container finished" podID="e1ef69eb-5928-4466-988d-6d4955721b2e" containerID="0039ba3c794fb38108bbe1bfd7777e5cbc69e11c18f20191cd06bb85b36515c5" exitCode=0 Jan 29 15:20:49 crc kubenswrapper[4767]: I0129 15:20:49.543159 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tst6c" event={"ID":"e1ef69eb-5928-4466-988d-6d4955721b2e","Type":"ContainerDied","Data":"0039ba3c794fb38108bbe1bfd7777e5cbc69e11c18f20191cd06bb85b36515c5"} Jan 29 15:20:50 crc kubenswrapper[4767]: I0129 15:20:50.567976 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zn5dm" event={"ID":"149720b9-74d1-432e-ad66-7824a18ff8a5","Type":"ContainerStarted","Data":"3779aa06323286222b45e6d0e4fe4e04846be116027b7006ff1cf2cc349fa53b"} Jan 29 15:20:50 crc kubenswrapper[4767]: I0129 15:20:50.570140 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tst6c" event={"ID":"e1ef69eb-5928-4466-988d-6d4955721b2e","Type":"ContainerStarted","Data":"6f4f71881b36811fc0cc10ed9037f80a132bf0f1aac2fd8459d749e7ecae3f4f"} Jan 29 15:20:50 crc kubenswrapper[4767]: I0129 15:20:50.584285 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zn5dm" podStartSLOduration=3.200202526 podStartE2EDuration="5.584268176s" podCreationTimestamp="2026-01-29 15:20:45 +0000 UTC" firstStartedPulling="2026-01-29 15:20:47.513537005 +0000 UTC m=+1203.323854044" lastFinishedPulling="2026-01-29 15:20:49.897602655 +0000 UTC m=+1205.707919694" observedRunningTime="2026-01-29 15:20:50.582776523 +0000 UTC m=+1206.393093572" watchObservedRunningTime="2026-01-29 15:20:50.584268176 +0000 UTC m=+1206.394585215" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:53.570869 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:53.570949 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:53.775936 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:53.776008 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:54.015793 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:54.042878 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tst6c" podStartSLOduration=6.531877885 podStartE2EDuration="9.042841525s" podCreationTimestamp="2026-01-29 15:20:45 +0000 UTC" firstStartedPulling="2026-01-29 15:20:47.507530633 +0000 UTC m=+1203.317847672" lastFinishedPulling="2026-01-29 15:20:50.018494273 +0000 UTC m=+1205.828811312" observedRunningTime="2026-01-29 15:20:50.598370701 +0000 UTC m=+1206.408687740" watchObservedRunningTime="2026-01-29 15:20:54.042841525 +0000 UTC m=+1209.853158564" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:54.062662 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c9h8z" Jan 29 15:20:54 crc kubenswrapper[4767]: I0129 15:20:54.823280 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sp2xf" podUID="e635388d-4adc-4dfe-aca6-4ed3c7ecc646" containerName="registry-server" probeResult="failure" output=< Jan 29 15:20:54 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:20:54 crc kubenswrapper[4767]: > Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.021349 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.021415 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.069606 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.192185 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.192267 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.231600 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.639683 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tst6c" Jan 29 15:20:56 crc kubenswrapper[4767]: I0129 15:20:56.647721 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zn5dm" Jan 29 15:21:03 crc kubenswrapper[4767]: I0129 15:21:03.830315 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:21:03 crc kubenswrapper[4767]: I0129 15:21:03.877480 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sp2xf" Jan 29 15:21:12 crc kubenswrapper[4767]: I0129 15:21:12.805666 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:21:12 crc kubenswrapper[4767]: I0129 15:21:12.806289 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.397284 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g"] Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.398860 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.403525 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.408013 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g"] Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.522613 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.522678 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.522747 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stvhj\" (UniqueName: \"kubernetes.io/projected/c36ab79e-316a-41b0-94fa-387a7fa88d5c-kube-api-access-stvhj\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.624367 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stvhj\" (UniqueName: \"kubernetes.io/projected/c36ab79e-316a-41b0-94fa-387a7fa88d5c-kube-api-access-stvhj\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.624447 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.624465 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.624859 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.625223 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.655022 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stvhj\" (UniqueName: \"kubernetes.io/projected/c36ab79e-316a-41b0-94fa-387a7fa88d5c-kube-api-access-stvhj\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.714636 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:15 crc kubenswrapper[4767]: I0129 15:21:15.896247 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g"] Jan 29 15:21:16 crc kubenswrapper[4767]: I0129 15:21:16.715534 4767 generic.go:334] "Generic (PLEG): container finished" podID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerID="8ffa63306c273d212a525c375ddc78357e7639c47628b401b107d90b60cf95e6" exitCode=0 Jan 29 15:21:16 crc kubenswrapper[4767]: I0129 15:21:16.715624 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" event={"ID":"c36ab79e-316a-41b0-94fa-387a7fa88d5c","Type":"ContainerDied","Data":"8ffa63306c273d212a525c375ddc78357e7639c47628b401b107d90b60cf95e6"} Jan 29 15:21:16 crc kubenswrapper[4767]: I0129 15:21:16.715818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" event={"ID":"c36ab79e-316a-41b0-94fa-387a7fa88d5c","Type":"ContainerStarted","Data":"51cf2e422a8c5b2ab6c2d78b6b3e057978c13e3f194ba9acd2f63c346b23cced"} Jan 29 15:21:18 crc kubenswrapper[4767]: I0129 15:21:18.730299 4767 generic.go:334] "Generic (PLEG): container finished" podID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerID="e06bcad826584943e2e18dde82c83f15f7ed5743cb3bf1359988ee80cdbdc782" exitCode=0 Jan 29 15:21:18 crc kubenswrapper[4767]: I0129 15:21:18.730408 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" event={"ID":"c36ab79e-316a-41b0-94fa-387a7fa88d5c","Type":"ContainerDied","Data":"e06bcad826584943e2e18dde82c83f15f7ed5743cb3bf1359988ee80cdbdc782"} Jan 29 15:21:19 crc kubenswrapper[4767]: I0129 15:21:19.739161 4767 generic.go:334] "Generic (PLEG): container finished" podID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerID="65faffd8fc21a61b568bd9eaf4c113cf137f5a3f130486138a77da0ce4595f46" exitCode=0 Jan 29 15:21:19 crc kubenswrapper[4767]: I0129 15:21:19.739208 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" event={"ID":"c36ab79e-316a-41b0-94fa-387a7fa88d5c","Type":"ContainerDied","Data":"65faffd8fc21a61b568bd9eaf4c113cf137f5a3f130486138a77da0ce4595f46"} Jan 29 15:21:20 crc kubenswrapper[4767]: I0129 15:21:20.987357 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.094257 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-util\") pod \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.094333 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-bundle\") pod \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.094365 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stvhj\" (UniqueName: \"kubernetes.io/projected/c36ab79e-316a-41b0-94fa-387a7fa88d5c-kube-api-access-stvhj\") pod \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\" (UID: \"c36ab79e-316a-41b0-94fa-387a7fa88d5c\") " Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.095188 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-bundle" (OuterVolumeSpecName: "bundle") pod "c36ab79e-316a-41b0-94fa-387a7fa88d5c" (UID: "c36ab79e-316a-41b0-94fa-387a7fa88d5c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.100875 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c36ab79e-316a-41b0-94fa-387a7fa88d5c-kube-api-access-stvhj" (OuterVolumeSpecName: "kube-api-access-stvhj") pod "c36ab79e-316a-41b0-94fa-387a7fa88d5c" (UID: "c36ab79e-316a-41b0-94fa-387a7fa88d5c"). InnerVolumeSpecName "kube-api-access-stvhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.111086 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-util" (OuterVolumeSpecName: "util") pod "c36ab79e-316a-41b0-94fa-387a7fa88d5c" (UID: "c36ab79e-316a-41b0-94fa-387a7fa88d5c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.196862 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-util\") on node \"crc\" DevicePath \"\"" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.196917 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c36ab79e-316a-41b0-94fa-387a7fa88d5c-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.196928 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stvhj\" (UniqueName: \"kubernetes.io/projected/c36ab79e-316a-41b0-94fa-387a7fa88d5c-kube-api-access-stvhj\") on node \"crc\" DevicePath \"\"" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.752938 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" event={"ID":"c36ab79e-316a-41b0-94fa-387a7fa88d5c","Type":"ContainerDied","Data":"51cf2e422a8c5b2ab6c2d78b6b3e057978c13e3f194ba9acd2f63c346b23cced"} Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.752990 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51cf2e422a8c5b2ab6c2d78b6b3e057978c13e3f194ba9acd2f63c346b23cced" Jan 29 15:21:21 crc kubenswrapper[4767]: I0129 15:21:21.753071 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.011995 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-9bjzl"] Jan 29 15:21:24 crc kubenswrapper[4767]: E0129 15:21:24.012952 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerName="util" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.013026 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerName="util" Jan 29 15:21:24 crc kubenswrapper[4767]: E0129 15:21:24.013093 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerName="pull" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.013145 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerName="pull" Jan 29 15:21:24 crc kubenswrapper[4767]: E0129 15:21:24.013198 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerName="extract" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.013249 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerName="extract" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.013411 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c36ab79e-316a-41b0-94fa-387a7fa88d5c" containerName="extract" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.013814 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.016347 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.017975 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.030246 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-9bjzl"] Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.032439 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-sn9gq" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.137165 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8v8m\" (UniqueName: \"kubernetes.io/projected/1a87ddd0-f27d-4331-9e42-b13d173825bb-kube-api-access-n8v8m\") pod \"nmstate-operator-646758c888-9bjzl\" (UID: \"1a87ddd0-f27d-4331-9e42-b13d173825bb\") " pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.239332 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8v8m\" (UniqueName: \"kubernetes.io/projected/1a87ddd0-f27d-4331-9e42-b13d173825bb-kube-api-access-n8v8m\") pod \"nmstate-operator-646758c888-9bjzl\" (UID: \"1a87ddd0-f27d-4331-9e42-b13d173825bb\") " pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.260386 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8v8m\" (UniqueName: \"kubernetes.io/projected/1a87ddd0-f27d-4331-9e42-b13d173825bb-kube-api-access-n8v8m\") pod \"nmstate-operator-646758c888-9bjzl\" (UID: \"1a87ddd0-f27d-4331-9e42-b13d173825bb\") " pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.336504 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.758411 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-9bjzl"] Jan 29 15:21:24 crc kubenswrapper[4767]: I0129 15:21:24.768682 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" event={"ID":"1a87ddd0-f27d-4331-9e42-b13d173825bb","Type":"ContainerStarted","Data":"99d15415717bfbf8307f7912c2309b40ec13e625e7f21184475e46f48b117b6b"} Jan 29 15:21:27 crc kubenswrapper[4767]: I0129 15:21:27.792156 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" event={"ID":"1a87ddd0-f27d-4331-9e42-b13d173825bb","Type":"ContainerStarted","Data":"07dbfa2dc306649c1a07be42aa876f4aa075691779b1d512b247605a36dceff1"} Jan 29 15:21:27 crc kubenswrapper[4767]: I0129 15:21:27.819640 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-9bjzl" podStartSLOduration=2.775498357 podStartE2EDuration="4.819608227s" podCreationTimestamp="2026-01-29 15:21:23 +0000 UTC" firstStartedPulling="2026-01-29 15:21:24.763534782 +0000 UTC m=+1240.573851821" lastFinishedPulling="2026-01-29 15:21:26.807644652 +0000 UTC m=+1242.617961691" observedRunningTime="2026-01-29 15:21:27.811264328 +0000 UTC m=+1243.621581367" watchObservedRunningTime="2026-01-29 15:21:27.819608227 +0000 UTC m=+1243.629925276" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.698339 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-5dlzw"] Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.699714 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.701671 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-v6798" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.706682 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf"] Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.707537 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.709781 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.722154 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-7pnbq"] Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.722931 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.726920 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-5dlzw"] Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.731806 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf"] Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.800256 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg9ws\" (UniqueName: \"kubernetes.io/projected/59099fce-32f9-4fc6-8ce4-e7789c00f336-kube-api-access-bg9ws\") pod \"nmstate-metrics-54757c584b-5dlzw\" (UID: \"59099fce-32f9-4fc6-8ce4-e7789c00f336\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.800542 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjnnd\" (UniqueName: \"kubernetes.io/projected/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-kube-api-access-vjnnd\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.800584 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-dbus-socket\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.800620 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-ovs-socket\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.800660 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42nf2\" (UniqueName: \"kubernetes.io/projected/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-kube-api-access-42nf2\") pod \"nmstate-webhook-8474b5b9d8-p4bhf\" (UID: \"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.800793 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-nmstate-lock\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.800883 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-p4bhf\" (UID: \"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.840034 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg"] Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.840659 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.843053 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.843648 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.845452 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-2bz5x" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.852855 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg"] Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902266 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-dbus-socket\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902382 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzj8t\" (UniqueName: \"kubernetes.io/projected/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-kube-api-access-rzj8t\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902431 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-ovs-socket\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902516 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-ovs-socket\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902559 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42nf2\" (UniqueName: \"kubernetes.io/projected/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-kube-api-access-42nf2\") pod \"nmstate-webhook-8474b5b9d8-p4bhf\" (UID: \"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902610 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902611 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-dbus-socket\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902741 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-nmstate-lock\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902819 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-p4bhf\" (UID: \"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902960 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:28 crc kubenswrapper[4767]: E0129 15:21:28.902991 4767 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.902958 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-nmstate-lock\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: E0129 15:21:28.903051 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-tls-key-pair podName:6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea nodeName:}" failed. No retries permitted until 2026-01-29 15:21:29.403028642 +0000 UTC m=+1245.213345751 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-tls-key-pair") pod "nmstate-webhook-8474b5b9d8-p4bhf" (UID: "6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea") : secret "openshift-nmstate-webhook" not found Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.903286 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg9ws\" (UniqueName: \"kubernetes.io/projected/59099fce-32f9-4fc6-8ce4-e7789c00f336-kube-api-access-bg9ws\") pod \"nmstate-metrics-54757c584b-5dlzw\" (UID: \"59099fce-32f9-4fc6-8ce4-e7789c00f336\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.903318 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjnnd\" (UniqueName: \"kubernetes.io/projected/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-kube-api-access-vjnnd\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.921448 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42nf2\" (UniqueName: \"kubernetes.io/projected/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-kube-api-access-42nf2\") pod \"nmstate-webhook-8474b5b9d8-p4bhf\" (UID: \"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.921971 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg9ws\" (UniqueName: \"kubernetes.io/projected/59099fce-32f9-4fc6-8ce4-e7789c00f336-kube-api-access-bg9ws\") pod \"nmstate-metrics-54757c584b-5dlzw\" (UID: \"59099fce-32f9-4fc6-8ce4-e7789c00f336\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" Jan 29 15:21:28 crc kubenswrapper[4767]: I0129 15:21:28.925121 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjnnd\" (UniqueName: \"kubernetes.io/projected/bc27129c-3bf6-4aeb-8a26-b46a555fe71f-kube-api-access-vjnnd\") pod \"nmstate-handler-7pnbq\" (UID: \"bc27129c-3bf6-4aeb-8a26-b46a555fe71f\") " pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.004357 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.004464 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzj8t\" (UniqueName: \"kubernetes.io/projected/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-kube-api-access-rzj8t\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.004516 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: E0129 15:21:29.004675 4767 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Jan 29 15:21:29 crc kubenswrapper[4767]: E0129 15:21:29.004738 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-plugin-serving-cert podName:cdbac033-2b2b-4f8a-a738-7304d0b0ac41 nodeName:}" failed. No retries permitted until 2026-01-29 15:21:29.504718208 +0000 UTC m=+1245.315035237 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-plugin-serving-cert") pod "nmstate-console-plugin-7754f76f8b-mrsvg" (UID: "cdbac033-2b2b-4f8a-a738-7304d0b0ac41") : secret "plugin-serving-cert" not found Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.005494 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.014949 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.033720 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzj8t\" (UniqueName: \"kubernetes.io/projected/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-kube-api-access-rzj8t\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.045210 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.161450 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5f88664d78-dbpsv"] Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.166281 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.180777 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f88664d78-dbpsv"] Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.206291 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw2j7\" (UniqueName: \"kubernetes.io/projected/8545cdec-dc54-49fd-b0c9-527b2330267d-kube-api-access-sw2j7\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.206333 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8545cdec-dc54-49fd-b0c9-527b2330267d-console-serving-cert\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.206350 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-console-config\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.206368 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-oauth-serving-cert\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.206415 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8545cdec-dc54-49fd-b0c9-527b2330267d-console-oauth-config\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.206476 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-service-ca\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.206491 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-trusted-ca-bundle\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.300089 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-5dlzw"] Jan 29 15:21:29 crc kubenswrapper[4767]: W0129 15:21:29.303198 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59099fce_32f9_4fc6_8ce4_e7789c00f336.slice/crio-fd55ccdfada3eca746f723ddb902e6e8f1c11e5425bb7a780fe2373e450547fb WatchSource:0}: Error finding container fd55ccdfada3eca746f723ddb902e6e8f1c11e5425bb7a780fe2373e450547fb: Status 404 returned error can't find the container with id fd55ccdfada3eca746f723ddb902e6e8f1c11e5425bb7a780fe2373e450547fb Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.308058 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8545cdec-dc54-49fd-b0c9-527b2330267d-console-oauth-config\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.308161 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-service-ca\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.308182 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-trusted-ca-bundle\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.308206 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw2j7\" (UniqueName: \"kubernetes.io/projected/8545cdec-dc54-49fd-b0c9-527b2330267d-kube-api-access-sw2j7\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.308234 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8545cdec-dc54-49fd-b0c9-527b2330267d-console-serving-cert\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.308254 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-console-config\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.308269 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-oauth-serving-cert\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.309631 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-oauth-serving-cert\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.311604 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-trusted-ca-bundle\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.312269 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-service-ca\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.312973 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8545cdec-dc54-49fd-b0c9-527b2330267d-console-config\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.321613 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8545cdec-dc54-49fd-b0c9-527b2330267d-console-serving-cert\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.321625 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8545cdec-dc54-49fd-b0c9-527b2330267d-console-oauth-config\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.328550 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw2j7\" (UniqueName: \"kubernetes.io/projected/8545cdec-dc54-49fd-b0c9-527b2330267d-kube-api-access-sw2j7\") pod \"console-5f88664d78-dbpsv\" (UID: \"8545cdec-dc54-49fd-b0c9-527b2330267d\") " pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.409605 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-p4bhf\" (UID: \"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.413529 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-p4bhf\" (UID: \"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.511760 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.516369 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/cdbac033-2b2b-4f8a-a738-7304d0b0ac41-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-mrsvg\" (UID: \"cdbac033-2b2b-4f8a-a738-7304d0b0ac41\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.522276 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.625981 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.705888 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5f88664d78-dbpsv"] Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.755456 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.802501 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7pnbq" event={"ID":"bc27129c-3bf6-4aeb-8a26-b46a555fe71f","Type":"ContainerStarted","Data":"945ee0af5b6ffabcbae81044bd98e6823941332cee957a8d5e72e3e23b02cca7"} Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.804063 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f88664d78-dbpsv" event={"ID":"8545cdec-dc54-49fd-b0c9-527b2330267d","Type":"ContainerStarted","Data":"139c71d61220a40d40f63eb91b85e21c0aa4dfc53c490f68e3299ef06fc90001"} Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.807437 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" event={"ID":"59099fce-32f9-4fc6-8ce4-e7789c00f336","Type":"ContainerStarted","Data":"fd55ccdfada3eca746f723ddb902e6e8f1c11e5425bb7a780fe2373e450547fb"} Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.820568 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf"] Jan 29 15:21:29 crc kubenswrapper[4767]: W0129 15:21:29.833521 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bcf1dba_60e6_4d8e_adb1_cd00eb9309ea.slice/crio-63bedaa5d91a2a2672b6df0b54dbadf4a14340648f9236f99d326a59e040fdc7 WatchSource:0}: Error finding container 63bedaa5d91a2a2672b6df0b54dbadf4a14340648f9236f99d326a59e040fdc7: Status 404 returned error can't find the container with id 63bedaa5d91a2a2672b6df0b54dbadf4a14340648f9236f99d326a59e040fdc7 Jan 29 15:21:29 crc kubenswrapper[4767]: I0129 15:21:29.949879 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg"] Jan 29 15:21:30 crc kubenswrapper[4767]: I0129 15:21:30.815227 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" event={"ID":"cdbac033-2b2b-4f8a-a738-7304d0b0ac41","Type":"ContainerStarted","Data":"a3b6c6784489c5c2294ba3e3756cb3689a5064c77de38b38c4d895eb9b829a93"} Jan 29 15:21:30 crc kubenswrapper[4767]: I0129 15:21:30.816583 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" event={"ID":"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea","Type":"ContainerStarted","Data":"63bedaa5d91a2a2672b6df0b54dbadf4a14340648f9236f99d326a59e040fdc7"} Jan 29 15:21:30 crc kubenswrapper[4767]: I0129 15:21:30.818461 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5f88664d78-dbpsv" event={"ID":"8545cdec-dc54-49fd-b0c9-527b2330267d","Type":"ContainerStarted","Data":"f9c7dce6c9083eebf6ac83764585083f175a62d507a604898e146ac45cb019b3"} Jan 29 15:21:30 crc kubenswrapper[4767]: I0129 15:21:30.842475 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5f88664d78-dbpsv" podStartSLOduration=1.8424546990000001 podStartE2EDuration="1.842454699s" podCreationTimestamp="2026-01-29 15:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:21:30.841163292 +0000 UTC m=+1246.651480351" watchObservedRunningTime="2026-01-29 15:21:30.842454699 +0000 UTC m=+1246.652771748" Jan 29 15:21:32 crc kubenswrapper[4767]: I0129 15:21:32.837100 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" event={"ID":"59099fce-32f9-4fc6-8ce4-e7789c00f336","Type":"ContainerStarted","Data":"34c4c7ec2be08a8e109948600c914eb9cfebfabe42c80c5b49943f857f9d2454"} Jan 29 15:21:32 crc kubenswrapper[4767]: I0129 15:21:32.839568 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" event={"ID":"6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea","Type":"ContainerStarted","Data":"51d86d96d2089bbfabc0bb261fdd6bb348ebe69001461a46697e8566e796991c"} Jan 29 15:21:32 crc kubenswrapper[4767]: I0129 15:21:32.839728 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:21:32 crc kubenswrapper[4767]: I0129 15:21:32.841666 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-7pnbq" event={"ID":"bc27129c-3bf6-4aeb-8a26-b46a555fe71f","Type":"ContainerStarted","Data":"174baf8d624065370a20cfbdb0e2ec859b8c0b7c914d36855bf21c54695a7305"} Jan 29 15:21:32 crc kubenswrapper[4767]: I0129 15:21:32.842099 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:32 crc kubenswrapper[4767]: I0129 15:21:32.862929 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" podStartSLOduration=2.8469603279999998 podStartE2EDuration="4.862884269s" podCreationTimestamp="2026-01-29 15:21:28 +0000 UTC" firstStartedPulling="2026-01-29 15:21:29.8359511 +0000 UTC m=+1245.646268139" lastFinishedPulling="2026-01-29 15:21:31.851875041 +0000 UTC m=+1247.662192080" observedRunningTime="2026-01-29 15:21:32.85768808 +0000 UTC m=+1248.668005109" watchObservedRunningTime="2026-01-29 15:21:32.862884269 +0000 UTC m=+1248.673201308" Jan 29 15:21:32 crc kubenswrapper[4767]: I0129 15:21:32.877923 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-7pnbq" podStartSLOduration=2.178447664 podStartE2EDuration="4.877808897s" podCreationTimestamp="2026-01-29 15:21:28 +0000 UTC" firstStartedPulling="2026-01-29 15:21:29.09820127 +0000 UTC m=+1244.908518309" lastFinishedPulling="2026-01-29 15:21:31.797562503 +0000 UTC m=+1247.607879542" observedRunningTime="2026-01-29 15:21:32.872758152 +0000 UTC m=+1248.683075191" watchObservedRunningTime="2026-01-29 15:21:32.877808897 +0000 UTC m=+1248.688125936" Jan 29 15:21:33 crc kubenswrapper[4767]: I0129 15:21:33.857868 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" event={"ID":"cdbac033-2b2b-4f8a-a738-7304d0b0ac41","Type":"ContainerStarted","Data":"3e564a24a4ee64fcbae728e81ef4e5a9577124c67dfec0632660d0983ffff999"} Jan 29 15:21:33 crc kubenswrapper[4767]: I0129 15:21:33.881047 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-mrsvg" podStartSLOduration=3.089269258 podStartE2EDuration="5.881018242s" podCreationTimestamp="2026-01-29 15:21:28 +0000 UTC" firstStartedPulling="2026-01-29 15:21:29.967569955 +0000 UTC m=+1245.777886994" lastFinishedPulling="2026-01-29 15:21:32.759318919 +0000 UTC m=+1248.569635978" observedRunningTime="2026-01-29 15:21:33.877187272 +0000 UTC m=+1249.687504311" watchObservedRunningTime="2026-01-29 15:21:33.881018242 +0000 UTC m=+1249.691335281" Jan 29 15:21:34 crc kubenswrapper[4767]: I0129 15:21:34.867494 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" event={"ID":"59099fce-32f9-4fc6-8ce4-e7789c00f336","Type":"ContainerStarted","Data":"38b8793a246c2484128d448cba98f820e1fcd92a2a5df7c5becda4964f85fd00"} Jan 29 15:21:34 crc kubenswrapper[4767]: I0129 15:21:34.886957 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-5dlzw" podStartSLOduration=2.39763497 podStartE2EDuration="6.886939014s" podCreationTimestamp="2026-01-29 15:21:28 +0000 UTC" firstStartedPulling="2026-01-29 15:21:29.305349001 +0000 UTC m=+1245.115666040" lastFinishedPulling="2026-01-29 15:21:33.794653045 +0000 UTC m=+1249.604970084" observedRunningTime="2026-01-29 15:21:34.886535813 +0000 UTC m=+1250.696852852" watchObservedRunningTime="2026-01-29 15:21:34.886939014 +0000 UTC m=+1250.697256053" Jan 29 15:21:39 crc kubenswrapper[4767]: I0129 15:21:39.075582 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-7pnbq" Jan 29 15:21:39 crc kubenswrapper[4767]: I0129 15:21:39.522637 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:39 crc kubenswrapper[4767]: I0129 15:21:39.522703 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:39 crc kubenswrapper[4767]: I0129 15:21:39.527560 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:39 crc kubenswrapper[4767]: I0129 15:21:39.909443 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5f88664d78-dbpsv" Jan 29 15:21:39 crc kubenswrapper[4767]: I0129 15:21:39.970056 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-89fkh"] Jan 29 15:21:42 crc kubenswrapper[4767]: I0129 15:21:42.806086 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:21:42 crc kubenswrapper[4767]: I0129 15:21:42.806416 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:21:42 crc kubenswrapper[4767]: I0129 15:21:42.806460 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:21:42 crc kubenswrapper[4767]: I0129 15:21:42.807103 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c74750ca0dbd8fc159be6c3f13ddce80420e211a2a5224b69e150664c60acb0e"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:21:42 crc kubenswrapper[4767]: I0129 15:21:42.807176 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://c74750ca0dbd8fc159be6c3f13ddce80420e211a2a5224b69e150664c60acb0e" gracePeriod=600 Jan 29 15:21:43 crc kubenswrapper[4767]: I0129 15:21:43.927278 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="c74750ca0dbd8fc159be6c3f13ddce80420e211a2a5224b69e150664c60acb0e" exitCode=0 Jan 29 15:21:43 crc kubenswrapper[4767]: I0129 15:21:43.927339 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"c74750ca0dbd8fc159be6c3f13ddce80420e211a2a5224b69e150664c60acb0e"} Jan 29 15:21:43 crc kubenswrapper[4767]: I0129 15:21:43.927606 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"20bafaf87627c386be9cca699d2328988c5557f20c2026d00c954ee5e4a7303e"} Jan 29 15:21:43 crc kubenswrapper[4767]: I0129 15:21:43.927626 4767 scope.go:117] "RemoveContainer" containerID="f49c65a1f42f7d0a539525b47266d6427aeb9e7ba92fb408c720097cf89b6158" Jan 29 15:21:49 crc kubenswrapper[4767]: I0129 15:21:49.631079 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-p4bhf" Jan 29 15:22:02 crc kubenswrapper[4767]: I0129 15:22:02.988533 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg"] Jan 29 15:22:02 crc kubenswrapper[4767]: I0129 15:22:02.990580 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:02 crc kubenswrapper[4767]: I0129 15:22:02.994309 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 29 15:22:02 crc kubenswrapper[4767]: I0129 15:22:02.998362 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg"] Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.180492 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6c49\" (UniqueName: \"kubernetes.io/projected/f5ab91f4-430d-45e9-b662-fc611e22e033-kube-api-access-m6c49\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.180575 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.180868 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.282522 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.282677 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.282709 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6c49\" (UniqueName: \"kubernetes.io/projected/f5ab91f4-430d-45e9-b662-fc611e22e033-kube-api-access-m6c49\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.283186 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.283210 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.303791 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6c49\" (UniqueName: \"kubernetes.io/projected/f5ab91f4-430d-45e9-b662-fc611e22e033-kube-api-access-m6c49\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.313216 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:03 crc kubenswrapper[4767]: I0129 15:22:03.515586 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg"] Jan 29 15:22:04 crc kubenswrapper[4767]: I0129 15:22:04.052112 4767 generic.go:334] "Generic (PLEG): container finished" podID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerID="bde9f86d69baca0776b5c8d740cc894180e46d3fe227e841d15f00bcada802b9" exitCode=0 Jan 29 15:22:04 crc kubenswrapper[4767]: I0129 15:22:04.052169 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" event={"ID":"f5ab91f4-430d-45e9-b662-fc611e22e033","Type":"ContainerDied","Data":"bde9f86d69baca0776b5c8d740cc894180e46d3fe227e841d15f00bcada802b9"} Jan 29 15:22:04 crc kubenswrapper[4767]: I0129 15:22:04.052457 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" event={"ID":"f5ab91f4-430d-45e9-b662-fc611e22e033","Type":"ContainerStarted","Data":"c30d9248f580773571af844fac0cbdd30a2d3db5acc98f2a23da09d519324d93"} Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.011163 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-89fkh" podUID="861eefc8-26d1-4726-857d-1e44816cdadb" containerName="console" containerID="cri-o://cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97" gracePeriod=15 Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.360213 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-89fkh_861eefc8-26d1-4726-857d-1e44816cdadb/console/0.log" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.360293 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.512056 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-oauth-config\") pod \"861eefc8-26d1-4726-857d-1e44816cdadb\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.512109 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-console-config\") pod \"861eefc8-26d1-4726-857d-1e44816cdadb\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.512149 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-oauth-serving-cert\") pod \"861eefc8-26d1-4726-857d-1e44816cdadb\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.512202 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgnbp\" (UniqueName: \"kubernetes.io/projected/861eefc8-26d1-4726-857d-1e44816cdadb-kube-api-access-pgnbp\") pod \"861eefc8-26d1-4726-857d-1e44816cdadb\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.512251 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-service-ca\") pod \"861eefc8-26d1-4726-857d-1e44816cdadb\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.512301 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-serving-cert\") pod \"861eefc8-26d1-4726-857d-1e44816cdadb\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.512331 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-trusted-ca-bundle\") pod \"861eefc8-26d1-4726-857d-1e44816cdadb\" (UID: \"861eefc8-26d1-4726-857d-1e44816cdadb\") " Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.513277 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-service-ca" (OuterVolumeSpecName: "service-ca") pod "861eefc8-26d1-4726-857d-1e44816cdadb" (UID: "861eefc8-26d1-4726-857d-1e44816cdadb"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.513313 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "861eefc8-26d1-4726-857d-1e44816cdadb" (UID: "861eefc8-26d1-4726-857d-1e44816cdadb"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.513323 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-console-config" (OuterVolumeSpecName: "console-config") pod "861eefc8-26d1-4726-857d-1e44816cdadb" (UID: "861eefc8-26d1-4726-857d-1e44816cdadb"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.513503 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "861eefc8-26d1-4726-857d-1e44816cdadb" (UID: "861eefc8-26d1-4726-857d-1e44816cdadb"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.518663 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "861eefc8-26d1-4726-857d-1e44816cdadb" (UID: "861eefc8-26d1-4726-857d-1e44816cdadb"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.519162 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "861eefc8-26d1-4726-857d-1e44816cdadb" (UID: "861eefc8-26d1-4726-857d-1e44816cdadb"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.519879 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/861eefc8-26d1-4726-857d-1e44816cdadb-kube-api-access-pgnbp" (OuterVolumeSpecName: "kube-api-access-pgnbp") pod "861eefc8-26d1-4726-857d-1e44816cdadb" (UID: "861eefc8-26d1-4726-857d-1e44816cdadb"). InnerVolumeSpecName "kube-api-access-pgnbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.614417 4767 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-service-ca\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.614738 4767 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.614755 4767 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.614768 4767 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/861eefc8-26d1-4726-857d-1e44816cdadb-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.614780 4767 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-console-config\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.614793 4767 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/861eefc8-26d1-4726-857d-1e44816cdadb-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:05 crc kubenswrapper[4767]: I0129 15:22:05.614806 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgnbp\" (UniqueName: \"kubernetes.io/projected/861eefc8-26d1-4726-857d-1e44816cdadb-kube-api-access-pgnbp\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.067532 4767 generic.go:334] "Generic (PLEG): container finished" podID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerID="6aa969387ddf7d61eeb0b02e78ac6d7ff0ac31d88cc88e2137623190fb87a00c" exitCode=0 Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.067644 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" event={"ID":"f5ab91f4-430d-45e9-b662-fc611e22e033","Type":"ContainerDied","Data":"6aa969387ddf7d61eeb0b02e78ac6d7ff0ac31d88cc88e2137623190fb87a00c"} Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.071067 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-89fkh_861eefc8-26d1-4726-857d-1e44816cdadb/console/0.log" Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.071107 4767 generic.go:334] "Generic (PLEG): container finished" podID="861eefc8-26d1-4726-857d-1e44816cdadb" containerID="cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97" exitCode=2 Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.071136 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-89fkh" event={"ID":"861eefc8-26d1-4726-857d-1e44816cdadb","Type":"ContainerDied","Data":"cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97"} Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.071168 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-89fkh" event={"ID":"861eefc8-26d1-4726-857d-1e44816cdadb","Type":"ContainerDied","Data":"48594a6f53da007ea41a8de4b1e1669fec04810c6b1705cd7f5bbfef0465ff3e"} Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.071172 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-89fkh" Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.071188 4767 scope.go:117] "RemoveContainer" containerID="cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97" Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.091113 4767 scope.go:117] "RemoveContainer" containerID="cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97" Jan 29 15:22:06 crc kubenswrapper[4767]: E0129 15:22:06.091573 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97\": container with ID starting with cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97 not found: ID does not exist" containerID="cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97" Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.091613 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97"} err="failed to get container status \"cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97\": rpc error: code = NotFound desc = could not find container \"cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97\": container with ID starting with cc826888ddd5da7b3bccc6732c362f3b133b38f4967b21717f5affa6d7f9bc97 not found: ID does not exist" Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.104984 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-89fkh"] Jan 29 15:22:06 crc kubenswrapper[4767]: I0129 15:22:06.123866 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-89fkh"] Jan 29 15:22:07 crc kubenswrapper[4767]: I0129 15:22:07.027180 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="861eefc8-26d1-4726-857d-1e44816cdadb" path="/var/lib/kubelet/pods/861eefc8-26d1-4726-857d-1e44816cdadb/volumes" Jan 29 15:22:07 crc kubenswrapper[4767]: I0129 15:22:07.082135 4767 generic.go:334] "Generic (PLEG): container finished" podID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerID="8a92b0f3728e221d97637e03fc2a44dcdc97d3c58bbeb898dbf7da5ca39f6e58" exitCode=0 Jan 29 15:22:07 crc kubenswrapper[4767]: I0129 15:22:07.082191 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" event={"ID":"f5ab91f4-430d-45e9-b662-fc611e22e033","Type":"ContainerDied","Data":"8a92b0f3728e221d97637e03fc2a44dcdc97d3c58bbeb898dbf7da5ca39f6e58"} Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.314827 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.455108 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-util\") pod \"f5ab91f4-430d-45e9-b662-fc611e22e033\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.455261 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-bundle\") pod \"f5ab91f4-430d-45e9-b662-fc611e22e033\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.455674 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6c49\" (UniqueName: \"kubernetes.io/projected/f5ab91f4-430d-45e9-b662-fc611e22e033-kube-api-access-m6c49\") pod \"f5ab91f4-430d-45e9-b662-fc611e22e033\" (UID: \"f5ab91f4-430d-45e9-b662-fc611e22e033\") " Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.456249 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-bundle" (OuterVolumeSpecName: "bundle") pod "f5ab91f4-430d-45e9-b662-fc611e22e033" (UID: "f5ab91f4-430d-45e9-b662-fc611e22e033"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.461108 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5ab91f4-430d-45e9-b662-fc611e22e033-kube-api-access-m6c49" (OuterVolumeSpecName: "kube-api-access-m6c49") pod "f5ab91f4-430d-45e9-b662-fc611e22e033" (UID: "f5ab91f4-430d-45e9-b662-fc611e22e033"). InnerVolumeSpecName "kube-api-access-m6c49". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.468619 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-util" (OuterVolumeSpecName: "util") pod "f5ab91f4-430d-45e9-b662-fc611e22e033" (UID: "f5ab91f4-430d-45e9-b662-fc611e22e033"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.557243 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-util\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.557303 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f5ab91f4-430d-45e9-b662-fc611e22e033-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:08 crc kubenswrapper[4767]: I0129 15:22:08.557318 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6c49\" (UniqueName: \"kubernetes.io/projected/f5ab91f4-430d-45e9-b662-fc611e22e033-kube-api-access-m6c49\") on node \"crc\" DevicePath \"\"" Jan 29 15:22:09 crc kubenswrapper[4767]: I0129 15:22:09.093685 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" event={"ID":"f5ab91f4-430d-45e9-b662-fc611e22e033","Type":"ContainerDied","Data":"c30d9248f580773571af844fac0cbdd30a2d3db5acc98f2a23da09d519324d93"} Jan 29 15:22:09 crc kubenswrapper[4767]: I0129 15:22:09.093786 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c30d9248f580773571af844fac0cbdd30a2d3db5acc98f2a23da09d519324d93" Jan 29 15:22:09 crc kubenswrapper[4767]: I0129 15:22:09.093732 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.243349 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk"] Jan 29 15:22:20 crc kubenswrapper[4767]: E0129 15:22:20.244112 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerName="util" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.244127 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerName="util" Jan 29 15:22:20 crc kubenswrapper[4767]: E0129 15:22:20.244142 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerName="pull" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.244149 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerName="pull" Jan 29 15:22:20 crc kubenswrapper[4767]: E0129 15:22:20.244161 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerName="extract" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.244168 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerName="extract" Jan 29 15:22:20 crc kubenswrapper[4767]: E0129 15:22:20.244185 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="861eefc8-26d1-4726-857d-1e44816cdadb" containerName="console" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.244192 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="861eefc8-26d1-4726-857d-1e44816cdadb" containerName="console" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.244304 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="861eefc8-26d1-4726-857d-1e44816cdadb" containerName="console" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.244323 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5ab91f4-430d-45e9-b662-fc611e22e033" containerName="extract" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.244753 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.246676 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.247559 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tg29w" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.247569 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.248024 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.249721 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.261498 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk"] Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.309698 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/80da1038-f75e-4ed8-9b53-dbd69e825d5e-webhook-cert\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.309752 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swwkp\" (UniqueName: \"kubernetes.io/projected/80da1038-f75e-4ed8-9b53-dbd69e825d5e-kube-api-access-swwkp\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.309833 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/80da1038-f75e-4ed8-9b53-dbd69e825d5e-apiservice-cert\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.412005 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/80da1038-f75e-4ed8-9b53-dbd69e825d5e-apiservice-cert\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.412120 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/80da1038-f75e-4ed8-9b53-dbd69e825d5e-webhook-cert\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.414201 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swwkp\" (UniqueName: \"kubernetes.io/projected/80da1038-f75e-4ed8-9b53-dbd69e825d5e-kube-api-access-swwkp\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.418056 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/80da1038-f75e-4ed8-9b53-dbd69e825d5e-apiservice-cert\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.418289 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/80da1038-f75e-4ed8-9b53-dbd69e825d5e-webhook-cert\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.434408 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swwkp\" (UniqueName: \"kubernetes.io/projected/80da1038-f75e-4ed8-9b53-dbd69e825d5e-kube-api-access-swwkp\") pod \"metallb-operator-controller-manager-875d98f97-bhdtk\" (UID: \"80da1038-f75e-4ed8-9b53-dbd69e825d5e\") " pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.478068 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb"] Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.478946 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.480441 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.481065 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.481843 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-gv4zk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.491421 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb"] Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.515633 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jgjp\" (UniqueName: \"kubernetes.io/projected/32caa2be-0b0c-4e06-b281-4f405ff743b6-kube-api-access-8jgjp\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.515709 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32caa2be-0b0c-4e06-b281-4f405ff743b6-webhook-cert\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.515742 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32caa2be-0b0c-4e06-b281-4f405ff743b6-apiservice-cert\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.559828 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.616402 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jgjp\" (UniqueName: \"kubernetes.io/projected/32caa2be-0b0c-4e06-b281-4f405ff743b6-kube-api-access-8jgjp\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.616481 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32caa2be-0b0c-4e06-b281-4f405ff743b6-webhook-cert\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.616509 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32caa2be-0b0c-4e06-b281-4f405ff743b6-apiservice-cert\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.622786 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/32caa2be-0b0c-4e06-b281-4f405ff743b6-apiservice-cert\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.628734 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/32caa2be-0b0c-4e06-b281-4f405ff743b6-webhook-cert\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.639176 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jgjp\" (UniqueName: \"kubernetes.io/projected/32caa2be-0b0c-4e06-b281-4f405ff743b6-kube-api-access-8jgjp\") pod \"metallb-operator-webhook-server-589648c7bd-vm5bb\" (UID: \"32caa2be-0b0c-4e06-b281-4f405ff743b6\") " pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.796217 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.985791 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb"] Jan 29 15:22:20 crc kubenswrapper[4767]: W0129 15:22:20.992026 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32caa2be_0b0c_4e06_b281_4f405ff743b6.slice/crio-cb50467d7f018252fff1e669cdeedcba8a58606e53c39d9bcefaed18d085f8fc WatchSource:0}: Error finding container cb50467d7f018252fff1e669cdeedcba8a58606e53c39d9bcefaed18d085f8fc: Status 404 returned error can't find the container with id cb50467d7f018252fff1e669cdeedcba8a58606e53c39d9bcefaed18d085f8fc Jan 29 15:22:20 crc kubenswrapper[4767]: I0129 15:22:20.997941 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk"] Jan 29 15:22:21 crc kubenswrapper[4767]: W0129 15:22:21.004109 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80da1038_f75e_4ed8_9b53_dbd69e825d5e.slice/crio-ec3e93e30038454f17ac136d612f8bf5e6970ab1f5a4e5339709fbd2c759abc8 WatchSource:0}: Error finding container ec3e93e30038454f17ac136d612f8bf5e6970ab1f5a4e5339709fbd2c759abc8: Status 404 returned error can't find the container with id ec3e93e30038454f17ac136d612f8bf5e6970ab1f5a4e5339709fbd2c759abc8 Jan 29 15:22:21 crc kubenswrapper[4767]: I0129 15:22:21.161216 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" event={"ID":"32caa2be-0b0c-4e06-b281-4f405ff743b6","Type":"ContainerStarted","Data":"cb50467d7f018252fff1e669cdeedcba8a58606e53c39d9bcefaed18d085f8fc"} Jan 29 15:22:21 crc kubenswrapper[4767]: I0129 15:22:21.162974 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" event={"ID":"80da1038-f75e-4ed8-9b53-dbd69e825d5e","Type":"ContainerStarted","Data":"ec3e93e30038454f17ac136d612f8bf5e6970ab1f5a4e5339709fbd2c759abc8"} Jan 29 15:22:25 crc kubenswrapper[4767]: I0129 15:22:25.193926 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" event={"ID":"80da1038-f75e-4ed8-9b53-dbd69e825d5e","Type":"ContainerStarted","Data":"a05d528313ce4e740a482598915872f2c087e3751854d4d09ee8eb3f508782ff"} Jan 29 15:22:25 crc kubenswrapper[4767]: I0129 15:22:25.194991 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:22:25 crc kubenswrapper[4767]: I0129 15:22:25.196458 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" event={"ID":"32caa2be-0b0c-4e06-b281-4f405ff743b6","Type":"ContainerStarted","Data":"d24e522cf54fcc203e00480379f5bfe321402bb0cc9abf523673a158bfb9060e"} Jan 29 15:22:25 crc kubenswrapper[4767]: I0129 15:22:25.196974 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:25 crc kubenswrapper[4767]: I0129 15:22:25.223275 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" podStartSLOduration=1.219126889 podStartE2EDuration="5.223256267s" podCreationTimestamp="2026-01-29 15:22:20 +0000 UTC" firstStartedPulling="2026-01-29 15:22:21.00852937 +0000 UTC m=+1296.818846409" lastFinishedPulling="2026-01-29 15:22:25.012658748 +0000 UTC m=+1300.822975787" observedRunningTime="2026-01-29 15:22:25.220289312 +0000 UTC m=+1301.030606351" watchObservedRunningTime="2026-01-29 15:22:25.223256267 +0000 UTC m=+1301.033573306" Jan 29 15:22:25 crc kubenswrapper[4767]: I0129 15:22:25.244601 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" podStartSLOduration=1.181014988 podStartE2EDuration="5.244578849s" podCreationTimestamp="2026-01-29 15:22:20 +0000 UTC" firstStartedPulling="2026-01-29 15:22:20.995473426 +0000 UTC m=+1296.805790465" lastFinishedPulling="2026-01-29 15:22:25.059037287 +0000 UTC m=+1300.869354326" observedRunningTime="2026-01-29 15:22:25.241330876 +0000 UTC m=+1301.051647915" watchObservedRunningTime="2026-01-29 15:22:25.244578849 +0000 UTC m=+1301.054895888" Jan 29 15:22:40 crc kubenswrapper[4767]: I0129 15:22:40.802601 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-589648c7bd-vm5bb" Jan 29 15:22:45 crc kubenswrapper[4767]: I0129 15:22:45.598727 4767 scope.go:117] "RemoveContainer" containerID="e48f5f18274bc43ddb52aec03a2c2150e11ba1b57213b4384e1335f3cef16973" Jan 29 15:22:45 crc kubenswrapper[4767]: I0129 15:22:45.622362 4767 scope.go:117] "RemoveContainer" containerID="d1f32304868de25c22dc62eb823bdbfa0939ea5d090761aa37af639fb2d9826e" Jan 29 15:23:00 crc kubenswrapper[4767]: I0129 15:23:00.561811 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-875d98f97-bhdtk" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.232045 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-6b5vt"] Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.234851 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.236647 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-dhjxg" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.236980 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k"] Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.237124 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.237691 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.237939 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.238825 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.257270 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k"] Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.322727 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-ldqv6"] Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.323580 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.327371 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.327645 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5ln8j" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.327758 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.330153 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.336250 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-h5qcl"] Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.337073 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.338290 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.338757 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vklt\" (UniqueName: \"kubernetes.io/projected/1e14bfd9-9551-45e2-85cd-75f27bc10ada-kube-api-access-8vklt\") pod \"frr-k8s-webhook-server-7df86c4f6c-qcf7k\" (UID: \"1e14bfd9-9551-45e2-85cd-75f27bc10ada\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.338917 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1e14bfd9-9551-45e2-85cd-75f27bc10ada-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-qcf7k\" (UID: \"1e14bfd9-9551-45e2-85cd-75f27bc10ada\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.359753 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-h5qcl"] Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.439785 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2jlg\" (UniqueName: \"kubernetes.io/projected/4304ecf7-d38c-4593-bd30-7e8de36c7638-kube-api-access-b2jlg\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.439841 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-conf\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.439876 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-reloader\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.439921 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vklt\" (UniqueName: \"kubernetes.io/projected/1e14bfd9-9551-45e2-85cd-75f27bc10ada-kube-api-access-8vklt\") pod \"frr-k8s-webhook-server-7df86c4f6c-qcf7k\" (UID: \"1e14bfd9-9551-45e2-85cd-75f27bc10ada\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.439946 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-memberlist\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.439971 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4304ecf7-d38c-4593-bd30-7e8de36c7638-metrics-certs\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.439997 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1e14bfd9-9551-45e2-85cd-75f27bc10ada-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-qcf7k\" (UID: \"1e14bfd9-9551-45e2-85cd-75f27bc10ada\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440022 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs4d5\" (UniqueName: \"kubernetes.io/projected/6ba067f0-0799-4fb8-ba5c-20745eaced80-kube-api-access-rs4d5\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440046 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6ba067f0-0799-4fb8-ba5c-20745eaced80-metallb-excludel2\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440066 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-cert\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440087 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-metrics-certs\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440114 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-startup\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440136 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-sockets\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440161 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxh8b\" (UniqueName: \"kubernetes.io/projected/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-kube-api-access-sxh8b\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440183 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-metrics-certs\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.440216 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-metrics\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: E0129 15:23:01.440613 4767 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Jan 29 15:23:01 crc kubenswrapper[4767]: E0129 15:23:01.440662 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1e14bfd9-9551-45e2-85cd-75f27bc10ada-cert podName:1e14bfd9-9551-45e2-85cd-75f27bc10ada nodeName:}" failed. No retries permitted until 2026-01-29 15:23:01.9406443 +0000 UTC m=+1337.750961339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1e14bfd9-9551-45e2-85cd-75f27bc10ada-cert") pod "frr-k8s-webhook-server-7df86c4f6c-qcf7k" (UID: "1e14bfd9-9551-45e2-85cd-75f27bc10ada") : secret "frr-k8s-webhook-server-cert" not found Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.456659 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vklt\" (UniqueName: \"kubernetes.io/projected/1e14bfd9-9551-45e2-85cd-75f27bc10ada-kube-api-access-8vklt\") pod \"frr-k8s-webhook-server-7df86c4f6c-qcf7k\" (UID: \"1e14bfd9-9551-45e2-85cd-75f27bc10ada\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.540943 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2jlg\" (UniqueName: \"kubernetes.io/projected/4304ecf7-d38c-4593-bd30-7e8de36c7638-kube-api-access-b2jlg\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541003 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-conf\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541037 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-reloader\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541060 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-memberlist\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541085 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4304ecf7-d38c-4593-bd30-7e8de36c7638-metrics-certs\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541129 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs4d5\" (UniqueName: \"kubernetes.io/projected/6ba067f0-0799-4fb8-ba5c-20745eaced80-kube-api-access-rs4d5\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541151 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6ba067f0-0799-4fb8-ba5c-20745eaced80-metallb-excludel2\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541170 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-cert\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541193 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-metrics-certs\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541220 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-startup\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541242 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-sockets\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541265 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxh8b\" (UniqueName: \"kubernetes.io/projected/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-kube-api-access-sxh8b\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541287 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-metrics-certs\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541318 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-metrics\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.541742 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-metrics\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.542258 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-conf\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.542476 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-reloader\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: E0129 15:23:01.542552 4767 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 29 15:23:01 crc kubenswrapper[4767]: E0129 15:23:01.542599 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-memberlist podName:6ba067f0-0799-4fb8-ba5c-20745eaced80 nodeName:}" failed. No retries permitted until 2026-01-29 15:23:02.042583104 +0000 UTC m=+1337.852900153 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-memberlist") pod "speaker-ldqv6" (UID: "6ba067f0-0799-4fb8-ba5c-20745eaced80") : secret "metallb-memberlist" not found Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.543861 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-startup\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.544606 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/6ba067f0-0799-4fb8-ba5c-20745eaced80-metallb-excludel2\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.547638 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4304ecf7-d38c-4593-bd30-7e8de36c7638-metrics-certs\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.548159 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4304ecf7-d38c-4593-bd30-7e8de36c7638-frr-sockets\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.555272 4767 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.555523 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-metrics-certs\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.555950 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-metrics-certs\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.566156 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxh8b\" (UniqueName: \"kubernetes.io/projected/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-kube-api-access-sxh8b\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.568164 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2jlg\" (UniqueName: \"kubernetes.io/projected/4304ecf7-d38c-4593-bd30-7e8de36c7638-kube-api-access-b2jlg\") pod \"frr-k8s-6b5vt\" (UID: \"4304ecf7-d38c-4593-bd30-7e8de36c7638\") " pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.568223 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd2bff9a-d96a-4e6d-b8b6-517b692c2c44-cert\") pod \"controller-6968d8fdc4-h5qcl\" (UID: \"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44\") " pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.588544 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs4d5\" (UniqueName: \"kubernetes.io/projected/6ba067f0-0799-4fb8-ba5c-20745eaced80-kube-api-access-rs4d5\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.657434 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.842237 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-h5qcl"] Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.855164 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.947659 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1e14bfd9-9551-45e2-85cd-75f27bc10ada-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-qcf7k\" (UID: \"1e14bfd9-9551-45e2-85cd-75f27bc10ada\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:01 crc kubenswrapper[4767]: I0129 15:23:01.954962 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1e14bfd9-9551-45e2-85cd-75f27bc10ada-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-qcf7k\" (UID: \"1e14bfd9-9551-45e2-85cd-75f27bc10ada\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.048988 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-memberlist\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.051731 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/6ba067f0-0799-4fb8-ba5c-20745eaced80-memberlist\") pod \"speaker-ldqv6\" (UID: \"6ba067f0-0799-4fb8-ba5c-20745eaced80\") " pod="metallb-system/speaker-ldqv6" Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.164765 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.247444 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-ldqv6" Jan 29 15:23:02 crc kubenswrapper[4767]: W0129 15:23:02.273399 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ba067f0_0799_4fb8_ba5c_20745eaced80.slice/crio-e4c962ae30f404bfcc6afacc5d4416f87bd4d7102baf545fd1ee13d6c58b7077 WatchSource:0}: Error finding container e4c962ae30f404bfcc6afacc5d4416f87bd4d7102baf545fd1ee13d6c58b7077: Status 404 returned error can't find the container with id e4c962ae30f404bfcc6afacc5d4416f87bd4d7102baf545fd1ee13d6c58b7077 Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.389031 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k"] Jan 29 15:23:02 crc kubenswrapper[4767]: W0129 15:23:02.395649 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e14bfd9_9551_45e2_85cd_75f27bc10ada.slice/crio-5d1ce3408d6073eb63b4abcdd7e966b39699eb6d41de580a65fd08c5440a067d WatchSource:0}: Error finding container 5d1ce3408d6073eb63b4abcdd7e966b39699eb6d41de580a65fd08c5440a067d: Status 404 returned error can't find the container with id 5d1ce3408d6073eb63b4abcdd7e966b39699eb6d41de580a65fd08c5440a067d Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.478747 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ldqv6" event={"ID":"6ba067f0-0799-4fb8-ba5c-20745eaced80","Type":"ContainerStarted","Data":"e4c962ae30f404bfcc6afacc5d4416f87bd4d7102baf545fd1ee13d6c58b7077"} Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.481069 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-h5qcl" event={"ID":"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44","Type":"ContainerStarted","Data":"2bef44cc31b1fbb99ea9cd34782a36bda93d95971f15dcd53b6fbf39bf9df5e0"} Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.481124 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-h5qcl" event={"ID":"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44","Type":"ContainerStarted","Data":"83f44332d9a0d4643f2f4cfa61720f3c091d75f732f72f617a30d169635a894d"} Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.481138 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-h5qcl" event={"ID":"dd2bff9a-d96a-4e6d-b8b6-517b692c2c44","Type":"ContainerStarted","Data":"330fcb10297a9e80a8065cacaf200af8ba93a5517d301a347a277035f1674258"} Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.481359 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.483698 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" event={"ID":"1e14bfd9-9551-45e2-85cd-75f27bc10ada","Type":"ContainerStarted","Data":"5d1ce3408d6073eb63b4abcdd7e966b39699eb6d41de580a65fd08c5440a067d"} Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.485965 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerStarted","Data":"8ed5782fe00a89b7490d32d599578fd75957521d370b182c2eb14e3336af842a"} Jan 29 15:23:02 crc kubenswrapper[4767]: I0129 15:23:02.503295 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-h5qcl" podStartSLOduration=1.503274159 podStartE2EDuration="1.503274159s" podCreationTimestamp="2026-01-29 15:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:23:02.494792895 +0000 UTC m=+1338.305109934" watchObservedRunningTime="2026-01-29 15:23:02.503274159 +0000 UTC m=+1338.313591198" Jan 29 15:23:03 crc kubenswrapper[4767]: I0129 15:23:03.494575 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ldqv6" event={"ID":"6ba067f0-0799-4fb8-ba5c-20745eaced80","Type":"ContainerStarted","Data":"1b3be99c70c41f965f067dad1229bcd96b9b485caabdf987ae7a3ccc255fa8ce"} Jan 29 15:23:03 crc kubenswrapper[4767]: I0129 15:23:03.494965 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-ldqv6" event={"ID":"6ba067f0-0799-4fb8-ba5c-20745eaced80","Type":"ContainerStarted","Data":"d62eb5853e411f35b66a8806b617d9bae98ab9dfab53171be4475fd32d821598"} Jan 29 15:23:03 crc kubenswrapper[4767]: I0129 15:23:03.519128 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-ldqv6" podStartSLOduration=2.519105415 podStartE2EDuration="2.519105415s" podCreationTimestamp="2026-01-29 15:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:23:03.515239214 +0000 UTC m=+1339.325556253" watchObservedRunningTime="2026-01-29 15:23:03.519105415 +0000 UTC m=+1339.329422464" Jan 29 15:23:04 crc kubenswrapper[4767]: I0129 15:23:04.514057 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-ldqv6" Jan 29 15:23:10 crc kubenswrapper[4767]: I0129 15:23:10.562823 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" event={"ID":"1e14bfd9-9551-45e2-85cd-75f27bc10ada","Type":"ContainerStarted","Data":"dd582ca11cc591c821ece9997eb49c097c7896942ce422b4815af53305661034"} Jan 29 15:23:10 crc kubenswrapper[4767]: I0129 15:23:10.565152 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:10 crc kubenswrapper[4767]: I0129 15:23:10.565417 4767 generic.go:334] "Generic (PLEG): container finished" podID="4304ecf7-d38c-4593-bd30-7e8de36c7638" containerID="8b807c26fc63645fb74cafbbdfb36a14e28afc8aaecf0aeb17af1b9f33be93d1" exitCode=0 Jan 29 15:23:10 crc kubenswrapper[4767]: I0129 15:23:10.565470 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerDied","Data":"8b807c26fc63645fb74cafbbdfb36a14e28afc8aaecf0aeb17af1b9f33be93d1"} Jan 29 15:23:10 crc kubenswrapper[4767]: I0129 15:23:10.588240 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" podStartSLOduration=2.559190714 podStartE2EDuration="9.588223472s" podCreationTimestamp="2026-01-29 15:23:01 +0000 UTC" firstStartedPulling="2026-01-29 15:23:02.397487094 +0000 UTC m=+1338.207804133" lastFinishedPulling="2026-01-29 15:23:09.426519852 +0000 UTC m=+1345.236836891" observedRunningTime="2026-01-29 15:23:10.582751195 +0000 UTC m=+1346.393068274" watchObservedRunningTime="2026-01-29 15:23:10.588223472 +0000 UTC m=+1346.398540511" Jan 29 15:23:11 crc kubenswrapper[4767]: I0129 15:23:11.572559 4767 generic.go:334] "Generic (PLEG): container finished" podID="4304ecf7-d38c-4593-bd30-7e8de36c7638" containerID="5cb7f26c21c6b1e83e023aaedb731aee39f34f39962c066ad889d8a9e25bac04" exitCode=0 Jan 29 15:23:11 crc kubenswrapper[4767]: I0129 15:23:11.572640 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerDied","Data":"5cb7f26c21c6b1e83e023aaedb731aee39f34f39962c066ad889d8a9e25bac04"} Jan 29 15:23:11 crc kubenswrapper[4767]: I0129 15:23:11.664245 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-h5qcl" Jan 29 15:23:12 crc kubenswrapper[4767]: I0129 15:23:12.251856 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-ldqv6" Jan 29 15:23:12 crc kubenswrapper[4767]: I0129 15:23:12.583775 4767 generic.go:334] "Generic (PLEG): container finished" podID="4304ecf7-d38c-4593-bd30-7e8de36c7638" containerID="4a11098ff515281a5605d0565ef4840ce50ec75454de4b2148b8a16e8583379f" exitCode=0 Jan 29 15:23:12 crc kubenswrapper[4767]: I0129 15:23:12.583882 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerDied","Data":"4a11098ff515281a5605d0565ef4840ce50ec75454de4b2148b8a16e8583379f"} Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.597734 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerStarted","Data":"ba7e396034e94f00a61743d9d5b02dcab228545a4fba43cfbac6c7434ee46030"} Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.598231 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.598248 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerStarted","Data":"7b7ecad5dac76cf2353b2db159583b7462f4601c5d859ad874116a67816ce15f"} Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.598261 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerStarted","Data":"bf27d8942551fd5ebb11daf8ac03a07273e5a2beb60e7d40e4d52cfd5600c582"} Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.598273 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerStarted","Data":"8c72c85a153895109cc9c3f7bd698542a95743d24636479af3e92c223a4241dc"} Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.598283 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerStarted","Data":"b90db2a38d34d2786f15c8d5595acba5f8e78f96189d53739635f1b905838a06"} Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.598294 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-6b5vt" event={"ID":"4304ecf7-d38c-4593-bd30-7e8de36c7638","Type":"ContainerStarted","Data":"45bddd43f1d29eba71cc3bc9487ef62367f044d166ebd80857a5b849f82dcbf8"} Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.622696 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7"] Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.624040 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.628518 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.629277 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-6b5vt" podStartSLOduration=5.155917225 podStartE2EDuration="12.629256686s" podCreationTimestamp="2026-01-29 15:23:01 +0000 UTC" firstStartedPulling="2026-01-29 15:23:01.972969918 +0000 UTC m=+1337.783286957" lastFinishedPulling="2026-01-29 15:23:09.446309379 +0000 UTC m=+1345.256626418" observedRunningTime="2026-01-29 15:23:13.624342475 +0000 UTC m=+1349.434659514" watchObservedRunningTime="2026-01-29 15:23:13.629256686 +0000 UTC m=+1349.439573725" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.635447 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7"] Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.715029 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.715083 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.715112 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q85sn\" (UniqueName: \"kubernetes.io/projected/e3d9aea9-f021-4d1a-866c-4ef98a702bae-kube-api-access-q85sn\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.816630 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.816674 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.816696 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q85sn\" (UniqueName: \"kubernetes.io/projected/e3d9aea9-f021-4d1a-866c-4ef98a702bae-kube-api-access-q85sn\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.817247 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.817309 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.835193 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q85sn\" (UniqueName: \"kubernetes.io/projected/e3d9aea9-f021-4d1a-866c-4ef98a702bae-kube-api-access-q85sn\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:13 crc kubenswrapper[4767]: I0129 15:23:13.966363 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:14 crc kubenswrapper[4767]: I0129 15:23:14.237792 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7"] Jan 29 15:23:14 crc kubenswrapper[4767]: I0129 15:23:14.603987 4767 generic.go:334] "Generic (PLEG): container finished" podID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerID="bf8f2180b60da8706e5e5dda0893b3defb08a63dabf036a89dc81a32bd448053" exitCode=0 Jan 29 15:23:14 crc kubenswrapper[4767]: I0129 15:23:14.604062 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" event={"ID":"e3d9aea9-f021-4d1a-866c-4ef98a702bae","Type":"ContainerDied","Data":"bf8f2180b60da8706e5e5dda0893b3defb08a63dabf036a89dc81a32bd448053"} Jan 29 15:23:14 crc kubenswrapper[4767]: I0129 15:23:14.604087 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" event={"ID":"e3d9aea9-f021-4d1a-866c-4ef98a702bae","Type":"ContainerStarted","Data":"ab8a110961310e151cc35ebbd0e5f98bce1d724f90aee3ab344286122627e588"} Jan 29 15:23:16 crc kubenswrapper[4767]: I0129 15:23:16.856234 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:16 crc kubenswrapper[4767]: I0129 15:23:16.920037 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:22 crc kubenswrapper[4767]: I0129 15:23:22.170106 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" Jan 29 15:23:25 crc kubenswrapper[4767]: I0129 15:23:25.686834 4767 generic.go:334] "Generic (PLEG): container finished" podID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerID="e223a33ea9ca7aa4f4733fad544f347893c013a58b83264aa5f36c41c7c21429" exitCode=0 Jan 29 15:23:25 crc kubenswrapper[4767]: I0129 15:23:25.687066 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" event={"ID":"e3d9aea9-f021-4d1a-866c-4ef98a702bae","Type":"ContainerDied","Data":"e223a33ea9ca7aa4f4733fad544f347893c013a58b83264aa5f36c41c7c21429"} Jan 29 15:23:26 crc kubenswrapper[4767]: I0129 15:23:26.697025 4767 generic.go:334] "Generic (PLEG): container finished" podID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerID="2ac8dd1c3ff90d9f852fe5de47a0c4d6e1227d6a07c5f4eb89aec1324cd6c798" exitCode=0 Jan 29 15:23:26 crc kubenswrapper[4767]: I0129 15:23:26.697087 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" event={"ID":"e3d9aea9-f021-4d1a-866c-4ef98a702bae","Type":"ContainerDied","Data":"2ac8dd1c3ff90d9f852fe5de47a0c4d6e1227d6a07c5f4eb89aec1324cd6c798"} Jan 29 15:23:27 crc kubenswrapper[4767]: I0129 15:23:27.935290 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.112922 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q85sn\" (UniqueName: \"kubernetes.io/projected/e3d9aea9-f021-4d1a-866c-4ef98a702bae-kube-api-access-q85sn\") pod \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.112982 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-util\") pod \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.113115 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-bundle\") pod \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\" (UID: \"e3d9aea9-f021-4d1a-866c-4ef98a702bae\") " Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.114430 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-bundle" (OuterVolumeSpecName: "bundle") pod "e3d9aea9-f021-4d1a-866c-4ef98a702bae" (UID: "e3d9aea9-f021-4d1a-866c-4ef98a702bae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.121065 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d9aea9-f021-4d1a-866c-4ef98a702bae-kube-api-access-q85sn" (OuterVolumeSpecName: "kube-api-access-q85sn") pod "e3d9aea9-f021-4d1a-866c-4ef98a702bae" (UID: "e3d9aea9-f021-4d1a-866c-4ef98a702bae"). InnerVolumeSpecName "kube-api-access-q85sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.122415 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-util" (OuterVolumeSpecName: "util") pod "e3d9aea9-f021-4d1a-866c-4ef98a702bae" (UID: "e3d9aea9-f021-4d1a-866c-4ef98a702bae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.214328 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.214413 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q85sn\" (UniqueName: \"kubernetes.io/projected/e3d9aea9-f021-4d1a-866c-4ef98a702bae-kube-api-access-q85sn\") on node \"crc\" DevicePath \"\"" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.214433 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3d9aea9-f021-4d1a-866c-4ef98a702bae-util\") on node \"crc\" DevicePath \"\"" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.712785 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" event={"ID":"e3d9aea9-f021-4d1a-866c-4ef98a702bae","Type":"ContainerDied","Data":"ab8a110961310e151cc35ebbd0e5f98bce1d724f90aee3ab344286122627e588"} Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.713180 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab8a110961310e151cc35ebbd0e5f98bce1d724f90aee3ab344286122627e588" Jan 29 15:23:28 crc kubenswrapper[4767]: I0129 15:23:28.712928 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7" Jan 29 15:23:31 crc kubenswrapper[4767]: I0129 15:23:31.860578 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-6b5vt" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.644315 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx"] Jan 29 15:23:37 crc kubenswrapper[4767]: E0129 15:23:37.644878 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerName="extract" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.644915 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerName="extract" Jan 29 15:23:37 crc kubenswrapper[4767]: E0129 15:23:37.644935 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerName="util" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.644945 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerName="util" Jan 29 15:23:37 crc kubenswrapper[4767]: E0129 15:23:37.644970 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerName="pull" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.644978 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerName="pull" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.645103 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d9aea9-f021-4d1a-866c-4ef98a702bae" containerName="extract" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.645610 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.647248 4767 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-5gb4r" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.647491 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.648938 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.666073 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx"] Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.725877 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a1db0450-30e0-4af7-b594-dd9855f96144-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-7vvmx\" (UID: \"a1db0450-30e0-4af7-b594-dd9855f96144\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.726001 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvjbz\" (UniqueName: \"kubernetes.io/projected/a1db0450-30e0-4af7-b594-dd9855f96144-kube-api-access-fvjbz\") pod \"cert-manager-operator-controller-manager-66c8bdd694-7vvmx\" (UID: \"a1db0450-30e0-4af7-b594-dd9855f96144\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.827018 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a1db0450-30e0-4af7-b594-dd9855f96144-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-7vvmx\" (UID: \"a1db0450-30e0-4af7-b594-dd9855f96144\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.827103 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvjbz\" (UniqueName: \"kubernetes.io/projected/a1db0450-30e0-4af7-b594-dd9855f96144-kube-api-access-fvjbz\") pod \"cert-manager-operator-controller-manager-66c8bdd694-7vvmx\" (UID: \"a1db0450-30e0-4af7-b594-dd9855f96144\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.827513 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a1db0450-30e0-4af7-b594-dd9855f96144-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-7vvmx\" (UID: \"a1db0450-30e0-4af7-b594-dd9855f96144\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.845036 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvjbz\" (UniqueName: \"kubernetes.io/projected/a1db0450-30e0-4af7-b594-dd9855f96144-kube-api-access-fvjbz\") pod \"cert-manager-operator-controller-manager-66c8bdd694-7vvmx\" (UID: \"a1db0450-30e0-4af7-b594-dd9855f96144\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:37 crc kubenswrapper[4767]: I0129 15:23:37.964395 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" Jan 29 15:23:38 crc kubenswrapper[4767]: I0129 15:23:38.386863 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx"] Jan 29 15:23:38 crc kubenswrapper[4767]: W0129 15:23:38.394156 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1db0450_30e0_4af7_b594_dd9855f96144.slice/crio-cbede10aa82a99f5bae13fb2d43b502bcf335fec7aac8f034684daeffe1958aa WatchSource:0}: Error finding container cbede10aa82a99f5bae13fb2d43b502bcf335fec7aac8f034684daeffe1958aa: Status 404 returned error can't find the container with id cbede10aa82a99f5bae13fb2d43b502bcf335fec7aac8f034684daeffe1958aa Jan 29 15:23:38 crc kubenswrapper[4767]: E0129 15:23:38.552149 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a" Jan 29 15:23:38 crc kubenswrapper[4767]: E0129 15:23:38.552522 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-operator,Image:registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a,Command:[/usr/bin/cert-manager-operator],Args:[start --v=$(OPERATOR_LOG_LEVEL) --trusted-ca-configmap=$(TRUSTED_CA_CONFIGMAP_NAME) --cloud-credentials-secret=$(CLOUD_CREDENTIALS_SECRET_NAME) --unsupported-addon-features=$(UNSUPPORTED_ADDON_FEATURES)],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cert-manager-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_WEBHOOK,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CA_INJECTOR,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CONTROLLER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ACMESOLVER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:38899dcd99bcd1c8c8d2c67cd19d5b1756434028ed2f1b926a282723bd63183e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ISTIOCSR,Value:registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:fb89adfcc4bcdaf21bed27dbc90586fb2b32180259b9f9d6fcf84a38e401fe03,ValueFrom:nil,},EnvVar{Name:OPERAND_IMAGE_VERSION,Value:1.18.4,ValueFrom:nil,},EnvVar{Name:ISTIOCSR_OPERAND_IMAGE_VERSION,Value:0.14.2,ValueFrom:nil,},EnvVar{Name:OPERATOR_IMAGE_VERSION,Value:1.18.1,ValueFrom:nil,},EnvVar{Name:OPERATOR_LOG_LEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:TRUSTED_CA_CONFIGMAP_NAME,Value:,ValueFrom:nil,},EnvVar{Name:CLOUD_CREDENTIALS_SECRET_NAME,Value:,ValueFrom:nil,},EnvVar{Name:UNSUPPORTED_ADDON_FEATURES,Value:,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cert-manager-operator.v1.18.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{33554432 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tmp,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fvjbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-operator-controller-manager-66c8bdd694-7vvmx_cert-manager-operator(a1db0450-30e0-4af7-b594-dd9855f96144): ErrImagePull: initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:23:38 crc kubenswrapper[4767]: E0129 15:23:38.554515 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ErrImagePull: \"initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:23:38 crc kubenswrapper[4767]: I0129 15:23:38.764120 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" event={"ID":"a1db0450-30e0-4af7-b594-dd9855f96144","Type":"ContainerStarted","Data":"cbede10aa82a99f5bae13fb2d43b502bcf335fec7aac8f034684daeffe1958aa"} Jan 29 15:23:38 crc kubenswrapper[4767]: E0129 15:23:38.765392 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:23:39 crc kubenswrapper[4767]: E0129 15:23:39.771888 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:23:50 crc kubenswrapper[4767]: E0129 15:23:50.155643 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a" Jan 29 15:23:50 crc kubenswrapper[4767]: E0129 15:23:50.156224 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-operator,Image:registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a,Command:[/usr/bin/cert-manager-operator],Args:[start --v=$(OPERATOR_LOG_LEVEL) --trusted-ca-configmap=$(TRUSTED_CA_CONFIGMAP_NAME) --cloud-credentials-secret=$(CLOUD_CREDENTIALS_SECRET_NAME) --unsupported-addon-features=$(UNSUPPORTED_ADDON_FEATURES)],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cert-manager-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_WEBHOOK,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CA_INJECTOR,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CONTROLLER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ACMESOLVER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:38899dcd99bcd1c8c8d2c67cd19d5b1756434028ed2f1b926a282723bd63183e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ISTIOCSR,Value:registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:fb89adfcc4bcdaf21bed27dbc90586fb2b32180259b9f9d6fcf84a38e401fe03,ValueFrom:nil,},EnvVar{Name:OPERAND_IMAGE_VERSION,Value:1.18.4,ValueFrom:nil,},EnvVar{Name:ISTIOCSR_OPERAND_IMAGE_VERSION,Value:0.14.2,ValueFrom:nil,},EnvVar{Name:OPERATOR_IMAGE_VERSION,Value:1.18.1,ValueFrom:nil,},EnvVar{Name:OPERATOR_LOG_LEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:TRUSTED_CA_CONFIGMAP_NAME,Value:,ValueFrom:nil,},EnvVar{Name:CLOUD_CREDENTIALS_SECRET_NAME,Value:,ValueFrom:nil,},EnvVar{Name:UNSUPPORTED_ADDON_FEATURES,Value:,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cert-manager-operator.v1.18.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{33554432 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tmp,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fvjbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-operator-controller-manager-66c8bdd694-7vvmx_cert-manager-operator(a1db0450-30e0-4af7-b594-dd9855f96144): ErrImagePull: initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:23:50 crc kubenswrapper[4767]: E0129 15:23:50.157387 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ErrImagePull: \"initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:24:04 crc kubenswrapper[4767]: E0129 15:24:04.020460 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:24:12 crc kubenswrapper[4767]: I0129 15:24:12.806753 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:24:12 crc kubenswrapper[4767]: I0129 15:24:12.807510 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:24:19 crc kubenswrapper[4767]: E0129 15:24:19.142787 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a" Jan 29 15:24:19 crc kubenswrapper[4767]: E0129 15:24:19.143677 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-operator,Image:registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a,Command:[/usr/bin/cert-manager-operator],Args:[start --v=$(OPERATOR_LOG_LEVEL) --trusted-ca-configmap=$(TRUSTED_CA_CONFIGMAP_NAME) --cloud-credentials-secret=$(CLOUD_CREDENTIALS_SECRET_NAME) --unsupported-addon-features=$(UNSUPPORTED_ADDON_FEATURES)],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cert-manager-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_WEBHOOK,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CA_INJECTOR,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CONTROLLER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ACMESOLVER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:38899dcd99bcd1c8c8d2c67cd19d5b1756434028ed2f1b926a282723bd63183e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ISTIOCSR,Value:registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:fb89adfcc4bcdaf21bed27dbc90586fb2b32180259b9f9d6fcf84a38e401fe03,ValueFrom:nil,},EnvVar{Name:OPERAND_IMAGE_VERSION,Value:1.18.4,ValueFrom:nil,},EnvVar{Name:ISTIOCSR_OPERAND_IMAGE_VERSION,Value:0.14.2,ValueFrom:nil,},EnvVar{Name:OPERATOR_IMAGE_VERSION,Value:1.18.1,ValueFrom:nil,},EnvVar{Name:OPERATOR_LOG_LEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:TRUSTED_CA_CONFIGMAP_NAME,Value:,ValueFrom:nil,},EnvVar{Name:CLOUD_CREDENTIALS_SECRET_NAME,Value:,ValueFrom:nil,},EnvVar{Name:UNSUPPORTED_ADDON_FEATURES,Value:,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cert-manager-operator.v1.18.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{33554432 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tmp,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fvjbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-operator-controller-manager-66c8bdd694-7vvmx_cert-manager-operator(a1db0450-30e0-4af7-b594-dd9855f96144): ErrImagePull: initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:24:19 crc kubenswrapper[4767]: E0129 15:24:19.144919 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ErrImagePull: \"initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:24:30 crc kubenswrapper[4767]: E0129 15:24:30.020769 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:24:42 crc kubenswrapper[4767]: I0129 15:24:42.809507 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:24:42 crc kubenswrapper[4767]: I0129 15:24:42.810028 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:24:43 crc kubenswrapper[4767]: E0129 15:24:43.021075 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:24:56 crc kubenswrapper[4767]: E0129 15:24:56.021799 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:25:11 crc kubenswrapper[4767]: E0129 15:25:11.146559 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a" Jan 29 15:25:11 crc kubenswrapper[4767]: E0129 15:25:11.147283 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-operator,Image:registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a,Command:[/usr/bin/cert-manager-operator],Args:[start --v=$(OPERATOR_LOG_LEVEL) --trusted-ca-configmap=$(TRUSTED_CA_CONFIGMAP_NAME) --cloud-credentials-secret=$(CLOUD_CREDENTIALS_SECRET_NAME) --unsupported-addon-features=$(UNSUPPORTED_ADDON_FEATURES)],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:cert-manager-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_WEBHOOK,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CA_INJECTOR,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_CONTROLLER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:903ce74138b1ffc735846a7c5fcdf62bbe82ca29568a6b38caec2656f6637671,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ACMESOLVER,Value:registry.redhat.io/cert-manager/jetstack-cert-manager-acmesolver-rhel9@sha256:38899dcd99bcd1c8c8d2c67cd19d5b1756434028ed2f1b926a282723bd63183e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CERT_MANAGER_ISTIOCSR,Value:registry.redhat.io/cert-manager/cert-manager-istio-csr-rhel9@sha256:fb89adfcc4bcdaf21bed27dbc90586fb2b32180259b9f9d6fcf84a38e401fe03,ValueFrom:nil,},EnvVar{Name:OPERAND_IMAGE_VERSION,Value:1.18.4,ValueFrom:nil,},EnvVar{Name:ISTIOCSR_OPERAND_IMAGE_VERSION,Value:0.14.2,ValueFrom:nil,},EnvVar{Name:OPERATOR_IMAGE_VERSION,Value:1.18.1,ValueFrom:nil,},EnvVar{Name:OPERATOR_LOG_LEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:TRUSTED_CA_CONFIGMAP_NAME,Value:,ValueFrom:nil,},EnvVar{Name:CLOUD_CREDENTIALS_SECRET_NAME,Value:,ValueFrom:nil,},EnvVar{Name:UNSUPPORTED_ADDON_FEATURES,Value:,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cert-manager-operator.v1.18.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{33554432 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:tmp,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fvjbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-operator-controller-manager-66c8bdd694-7vvmx_cert-manager-operator(a1db0450-30e0-4af7-b594-dd9855f96144): ErrImagePull: initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:25:11 crc kubenswrapper[4767]: E0129 15:25:11.148996 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ErrImagePull: \"initializing source docker://registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:25:12 crc kubenswrapper[4767]: I0129 15:25:12.805712 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:25:12 crc kubenswrapper[4767]: I0129 15:25:12.806115 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:25:12 crc kubenswrapper[4767]: I0129 15:25:12.806556 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:25:12 crc kubenswrapper[4767]: I0129 15:25:12.807502 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"20bafaf87627c386be9cca699d2328988c5557f20c2026d00c954ee5e4a7303e"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:25:12 crc kubenswrapper[4767]: I0129 15:25:12.807590 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://20bafaf87627c386be9cca699d2328988c5557f20c2026d00c954ee5e4a7303e" gracePeriod=600 Jan 29 15:25:13 crc kubenswrapper[4767]: I0129 15:25:13.406110 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="20bafaf87627c386be9cca699d2328988c5557f20c2026d00c954ee5e4a7303e" exitCode=0 Jan 29 15:25:13 crc kubenswrapper[4767]: I0129 15:25:13.406176 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"20bafaf87627c386be9cca699d2328988c5557f20c2026d00c954ee5e4a7303e"} Jan 29 15:25:13 crc kubenswrapper[4767]: I0129 15:25:13.406505 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce"} Jan 29 15:25:13 crc kubenswrapper[4767]: I0129 15:25:13.406553 4767 scope.go:117] "RemoveContainer" containerID="c74750ca0dbd8fc159be6c3f13ddce80420e211a2a5224b69e150664c60acb0e" Jan 29 15:25:24 crc kubenswrapper[4767]: E0129 15:25:24.021172 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:25:38 crc kubenswrapper[4767]: E0129 15:25:38.021423 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:25:49 crc kubenswrapper[4767]: E0129 15:25:49.020843 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:26:02 crc kubenswrapper[4767]: E0129 15:26:02.020778 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:26:15 crc kubenswrapper[4767]: E0129 15:26:15.024386 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:26:26 crc kubenswrapper[4767]: E0129 15:26:26.022861 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/cert-manager-operator-rhel9@sha256:99526f5a179816df1f7f51df0517136b247d815b7bdce0a5d0eb7cdaf4b5ce7a\\\"\"" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podUID="a1db0450-30e0-4af7-b594-dd9855f96144" Jan 29 15:26:40 crc kubenswrapper[4767]: I0129 15:26:40.019833 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:26:43 crc kubenswrapper[4767]: I0129 15:26:43.013120 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" event={"ID":"a1db0450-30e0-4af7-b594-dd9855f96144","Type":"ContainerStarted","Data":"d1e3ec8a2e5b580e0968f18d3574e95510a3782ec7ac400a4ce146af9a17eb59"} Jan 29 15:26:43 crc kubenswrapper[4767]: I0129 15:26:43.036295 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-7vvmx" podStartSLOduration=2.031274462 podStartE2EDuration="3m6.036279837s" podCreationTimestamp="2026-01-29 15:23:37 +0000 UTC" firstStartedPulling="2026-01-29 15:23:38.396581736 +0000 UTC m=+1374.206898775" lastFinishedPulling="2026-01-29 15:26:42.401587111 +0000 UTC m=+1558.211904150" observedRunningTime="2026-01-29 15:26:43.033469406 +0000 UTC m=+1558.843786445" watchObservedRunningTime="2026-01-29 15:26:43.036279837 +0000 UTC m=+1558.846596876" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.370955 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-hp7hd"] Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.371889 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.377641 4767 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-nvk52" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.377650 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.378929 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.390775 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-hp7hd"] Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.510569 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0649463b-73be-45e2-a8f3-393f0003d054-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-hp7hd\" (UID: \"0649463b-73be-45e2-a8f3-393f0003d054\") " pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.510619 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tk5n\" (UniqueName: \"kubernetes.io/projected/0649463b-73be-45e2-a8f3-393f0003d054-kube-api-access-7tk5n\") pod \"cert-manager-webhook-6888856db4-hp7hd\" (UID: \"0649463b-73be-45e2-a8f3-393f0003d054\") " pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.612230 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0649463b-73be-45e2-a8f3-393f0003d054-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-hp7hd\" (UID: \"0649463b-73be-45e2-a8f3-393f0003d054\") " pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.612319 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tk5n\" (UniqueName: \"kubernetes.io/projected/0649463b-73be-45e2-a8f3-393f0003d054-kube-api-access-7tk5n\") pod \"cert-manager-webhook-6888856db4-hp7hd\" (UID: \"0649463b-73be-45e2-a8f3-393f0003d054\") " pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.635354 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tk5n\" (UniqueName: \"kubernetes.io/projected/0649463b-73be-45e2-a8f3-393f0003d054-kube-api-access-7tk5n\") pod \"cert-manager-webhook-6888856db4-hp7hd\" (UID: \"0649463b-73be-45e2-a8f3-393f0003d054\") " pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.635495 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0649463b-73be-45e2-a8f3-393f0003d054-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-hp7hd\" (UID: \"0649463b-73be-45e2-a8f3-393f0003d054\") " pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:46 crc kubenswrapper[4767]: I0129 15:26:46.688622 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:47 crc kubenswrapper[4767]: I0129 15:26:47.158407 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-hp7hd"] Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.041812 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" event={"ID":"0649463b-73be-45e2-a8f3-393f0003d054","Type":"ContainerStarted","Data":"8803a654e099578874d502d88011cce6016d39e0020952199c4f5e1d4836bf80"} Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.527110 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-k46vw"] Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.528312 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.531151 4767 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-7h8tb" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.537376 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-k46vw"] Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.540779 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/66377cb0-a5db-4193-888c-8252e286f1a4-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-k46vw\" (UID: \"66377cb0-a5db-4193-888c-8252e286f1a4\") " pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.540847 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmg8n\" (UniqueName: \"kubernetes.io/projected/66377cb0-a5db-4193-888c-8252e286f1a4-kube-api-access-fmg8n\") pod \"cert-manager-cainjector-5545bd876-k46vw\" (UID: \"66377cb0-a5db-4193-888c-8252e286f1a4\") " pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.642684 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/66377cb0-a5db-4193-888c-8252e286f1a4-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-k46vw\" (UID: \"66377cb0-a5db-4193-888c-8252e286f1a4\") " pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.642756 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmg8n\" (UniqueName: \"kubernetes.io/projected/66377cb0-a5db-4193-888c-8252e286f1a4-kube-api-access-fmg8n\") pod \"cert-manager-cainjector-5545bd876-k46vw\" (UID: \"66377cb0-a5db-4193-888c-8252e286f1a4\") " pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.661641 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/66377cb0-a5db-4193-888c-8252e286f1a4-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-k46vw\" (UID: \"66377cb0-a5db-4193-888c-8252e286f1a4\") " pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.667630 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmg8n\" (UniqueName: \"kubernetes.io/projected/66377cb0-a5db-4193-888c-8252e286f1a4-kube-api-access-fmg8n\") pod \"cert-manager-cainjector-5545bd876-k46vw\" (UID: \"66377cb0-a5db-4193-888c-8252e286f1a4\") " pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:48 crc kubenswrapper[4767]: I0129 15:26:48.853310 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" Jan 29 15:26:49 crc kubenswrapper[4767]: I0129 15:26:49.280145 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-k46vw"] Jan 29 15:26:50 crc kubenswrapper[4767]: I0129 15:26:50.075543 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" event={"ID":"66377cb0-a5db-4193-888c-8252e286f1a4","Type":"ContainerStarted","Data":"68312e0e632c270e820dc514dcd5528e8e13a1dc1e82c3440d34812700260056"} Jan 29 15:26:52 crc kubenswrapper[4767]: I0129 15:26:52.088489 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" event={"ID":"66377cb0-a5db-4193-888c-8252e286f1a4","Type":"ContainerStarted","Data":"41c474e6ced9669acb0c36f3e42566c72aa63f7c6c2b7a9867d705a73d2c3ef9"} Jan 29 15:26:52 crc kubenswrapper[4767]: I0129 15:26:52.091010 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" event={"ID":"0649463b-73be-45e2-a8f3-393f0003d054","Type":"ContainerStarted","Data":"f9ec773d8e7c32c7ed3c2d8fe57bb098af78f79927c0d30ddf4f7bfd3d3c0ec4"} Jan 29 15:26:52 crc kubenswrapper[4767]: I0129 15:26:52.091149 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:52 crc kubenswrapper[4767]: I0129 15:26:52.112385 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-k46vw" podStartSLOduration=1.9685811050000002 podStartE2EDuration="4.112369369s" podCreationTimestamp="2026-01-29 15:26:48 +0000 UTC" firstStartedPulling="2026-01-29 15:26:49.284447033 +0000 UTC m=+1565.094764072" lastFinishedPulling="2026-01-29 15:26:51.428235297 +0000 UTC m=+1567.238552336" observedRunningTime="2026-01-29 15:26:52.10643867 +0000 UTC m=+1567.916755709" watchObservedRunningTime="2026-01-29 15:26:52.112369369 +0000 UTC m=+1567.922686408" Jan 29 15:26:52 crc kubenswrapper[4767]: I0129 15:26:52.133744 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" podStartSLOduration=1.875935605 podStartE2EDuration="6.133730671s" podCreationTimestamp="2026-01-29 15:26:46 +0000 UTC" firstStartedPulling="2026-01-29 15:26:47.16795593 +0000 UTC m=+1562.978272979" lastFinishedPulling="2026-01-29 15:26:51.425751006 +0000 UTC m=+1567.236068045" observedRunningTime="2026-01-29 15:26:52.130810318 +0000 UTC m=+1567.941127357" watchObservedRunningTime="2026-01-29 15:26:52.133730671 +0000 UTC m=+1567.944047700" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.034460 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-tztj2"] Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.035858 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.041548 4767 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-94c7d" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.042770 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-989s8\" (UniqueName: \"kubernetes.io/projected/e9862466-1c30-4644-bf55-8e9f1548555e-kube-api-access-989s8\") pod \"cert-manager-545d4d4674-tztj2\" (UID: \"e9862466-1c30-4644-bf55-8e9f1548555e\") " pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.043040 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9862466-1c30-4644-bf55-8e9f1548555e-bound-sa-token\") pod \"cert-manager-545d4d4674-tztj2\" (UID: \"e9862466-1c30-4644-bf55-8e9f1548555e\") " pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.044737 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-tztj2"] Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.144528 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9862466-1c30-4644-bf55-8e9f1548555e-bound-sa-token\") pod \"cert-manager-545d4d4674-tztj2\" (UID: \"e9862466-1c30-4644-bf55-8e9f1548555e\") " pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.144598 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-989s8\" (UniqueName: \"kubernetes.io/projected/e9862466-1c30-4644-bf55-8e9f1548555e-kube-api-access-989s8\") pod \"cert-manager-545d4d4674-tztj2\" (UID: \"e9862466-1c30-4644-bf55-8e9f1548555e\") " pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.169344 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9862466-1c30-4644-bf55-8e9f1548555e-bound-sa-token\") pod \"cert-manager-545d4d4674-tztj2\" (UID: \"e9862466-1c30-4644-bf55-8e9f1548555e\") " pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.172939 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-989s8\" (UniqueName: \"kubernetes.io/projected/e9862466-1c30-4644-bf55-8e9f1548555e-kube-api-access-989s8\") pod \"cert-manager-545d4d4674-tztj2\" (UID: \"e9862466-1c30-4644-bf55-8e9f1548555e\") " pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.357477 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-tztj2" Jan 29 15:26:55 crc kubenswrapper[4767]: I0129 15:26:55.583580 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-tztj2"] Jan 29 15:26:55 crc kubenswrapper[4767]: W0129 15:26:55.587572 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9862466_1c30_4644_bf55_8e9f1548555e.slice/crio-958c79728a23b5ccb0fc36986fc4cad9e7ed3aba0c03f1c416a213ec7295a334 WatchSource:0}: Error finding container 958c79728a23b5ccb0fc36986fc4cad9e7ed3aba0c03f1c416a213ec7295a334: Status 404 returned error can't find the container with id 958c79728a23b5ccb0fc36986fc4cad9e7ed3aba0c03f1c416a213ec7295a334 Jan 29 15:26:56 crc kubenswrapper[4767]: I0129 15:26:56.118193 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-tztj2" event={"ID":"e9862466-1c30-4644-bf55-8e9f1548555e","Type":"ContainerStarted","Data":"105d944e49d7264ac9bc4b441e765a78e7b9b10440ebde692259b6b749104a2e"} Jan 29 15:26:56 crc kubenswrapper[4767]: I0129 15:26:56.118266 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-tztj2" event={"ID":"e9862466-1c30-4644-bf55-8e9f1548555e","Type":"ContainerStarted","Data":"958c79728a23b5ccb0fc36986fc4cad9e7ed3aba0c03f1c416a213ec7295a334"} Jan 29 15:26:56 crc kubenswrapper[4767]: I0129 15:26:56.141210 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-tztj2" podStartSLOduration=1.141186787 podStartE2EDuration="1.141186787s" podCreationTimestamp="2026-01-29 15:26:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:26:56.138463759 +0000 UTC m=+1571.948780798" watchObservedRunningTime="2026-01-29 15:26:56.141186787 +0000 UTC m=+1571.951503826" Jan 29 15:26:56 crc kubenswrapper[4767]: I0129 15:26:56.691478 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-hp7hd" Jan 29 15:26:59 crc kubenswrapper[4767]: I0129 15:26:59.992833 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fl7z6"] Jan 29 15:26:59 crc kubenswrapper[4767]: I0129 15:26:59.994078 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fl7z6" Jan 29 15:26:59 crc kubenswrapper[4767]: I0129 15:26:59.996184 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 29 15:26:59 crc kubenswrapper[4767]: I0129 15:26:59.996184 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gz42h" Jan 29 15:26:59 crc kubenswrapper[4767]: I0129 15:26:59.999588 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 29 15:27:00 crc kubenswrapper[4767]: I0129 15:27:00.002536 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fl7z6"] Jan 29 15:27:00 crc kubenswrapper[4767]: I0129 15:27:00.113460 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpw6d\" (UniqueName: \"kubernetes.io/projected/8f46289b-9b6d-41df-aa0e-ad2b40216d0f-kube-api-access-bpw6d\") pod \"openstack-operator-index-fl7z6\" (UID: \"8f46289b-9b6d-41df-aa0e-ad2b40216d0f\") " pod="openstack-operators/openstack-operator-index-fl7z6" Jan 29 15:27:00 crc kubenswrapper[4767]: I0129 15:27:00.215198 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpw6d\" (UniqueName: \"kubernetes.io/projected/8f46289b-9b6d-41df-aa0e-ad2b40216d0f-kube-api-access-bpw6d\") pod \"openstack-operator-index-fl7z6\" (UID: \"8f46289b-9b6d-41df-aa0e-ad2b40216d0f\") " pod="openstack-operators/openstack-operator-index-fl7z6" Jan 29 15:27:00 crc kubenswrapper[4767]: I0129 15:27:00.238951 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpw6d\" (UniqueName: \"kubernetes.io/projected/8f46289b-9b6d-41df-aa0e-ad2b40216d0f-kube-api-access-bpw6d\") pod \"openstack-operator-index-fl7z6\" (UID: \"8f46289b-9b6d-41df-aa0e-ad2b40216d0f\") " pod="openstack-operators/openstack-operator-index-fl7z6" Jan 29 15:27:00 crc kubenswrapper[4767]: I0129 15:27:00.354165 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fl7z6" Jan 29 15:27:00 crc kubenswrapper[4767]: I0129 15:27:00.632452 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fl7z6"] Jan 29 15:27:00 crc kubenswrapper[4767]: W0129 15:27:00.634778 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f46289b_9b6d_41df_aa0e_ad2b40216d0f.slice/crio-345fbb32e484196941eacd475c47becc9955589c4d1dd92f75020bb2b0ac792f WatchSource:0}: Error finding container 345fbb32e484196941eacd475c47becc9955589c4d1dd92f75020bb2b0ac792f: Status 404 returned error can't find the container with id 345fbb32e484196941eacd475c47becc9955589c4d1dd92f75020bb2b0ac792f Jan 29 15:27:01 crc kubenswrapper[4767]: I0129 15:27:01.160025 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fl7z6" event={"ID":"8f46289b-9b6d-41df-aa0e-ad2b40216d0f","Type":"ContainerStarted","Data":"345fbb32e484196941eacd475c47becc9955589c4d1dd92f75020bb2b0ac792f"} Jan 29 15:27:02 crc kubenswrapper[4767]: I0129 15:27:02.182012 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fl7z6"] Jan 29 15:27:02 crc kubenswrapper[4767]: I0129 15:27:02.773281 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tsfnc"] Jan 29 15:27:02 crc kubenswrapper[4767]: I0129 15:27:02.774336 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:02 crc kubenswrapper[4767]: I0129 15:27:02.784150 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tsfnc"] Jan 29 15:27:02 crc kubenswrapper[4767]: I0129 15:27:02.952819 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j942q\" (UniqueName: \"kubernetes.io/projected/12bdee23-d478-434a-95bc-7d2c042f6906-kube-api-access-j942q\") pod \"openstack-operator-index-tsfnc\" (UID: \"12bdee23-d478-434a-95bc-7d2c042f6906\") " pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.054782 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j942q\" (UniqueName: \"kubernetes.io/projected/12bdee23-d478-434a-95bc-7d2c042f6906-kube-api-access-j942q\") pod \"openstack-operator-index-tsfnc\" (UID: \"12bdee23-d478-434a-95bc-7d2c042f6906\") " pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.081138 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j942q\" (UniqueName: \"kubernetes.io/projected/12bdee23-d478-434a-95bc-7d2c042f6906-kube-api-access-j942q\") pod \"openstack-operator-index-tsfnc\" (UID: \"12bdee23-d478-434a-95bc-7d2c042f6906\") " pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.095829 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.180877 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fl7z6" event={"ID":"8f46289b-9b6d-41df-aa0e-ad2b40216d0f","Type":"ContainerStarted","Data":"357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec"} Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.181027 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-fl7z6" podUID="8f46289b-9b6d-41df-aa0e-ad2b40216d0f" containerName="registry-server" containerID="cri-o://357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec" gracePeriod=2 Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.200000 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fl7z6" podStartSLOduration=1.909919304 podStartE2EDuration="4.199983098s" podCreationTimestamp="2026-01-29 15:26:59 +0000 UTC" firstStartedPulling="2026-01-29 15:27:00.641641561 +0000 UTC m=+1576.451958600" lastFinishedPulling="2026-01-29 15:27:02.931705355 +0000 UTC m=+1578.742022394" observedRunningTime="2026-01-29 15:27:03.199031011 +0000 UTC m=+1579.009348050" watchObservedRunningTime="2026-01-29 15:27:03.199983098 +0000 UTC m=+1579.010300137" Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.509527 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tsfnc"] Jan 29 15:27:03 crc kubenswrapper[4767]: W0129 15:27:03.515474 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12bdee23_d478_434a_95bc_7d2c042f6906.slice/crio-7cf19913ce6fc1768714b57bab3f8cbc9c5403cdf925dfb632bde38c31f58ef9 WatchSource:0}: Error finding container 7cf19913ce6fc1768714b57bab3f8cbc9c5403cdf925dfb632bde38c31f58ef9: Status 404 returned error can't find the container with id 7cf19913ce6fc1768714b57bab3f8cbc9c5403cdf925dfb632bde38c31f58ef9 Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.526494 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fl7z6" Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.661642 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpw6d\" (UniqueName: \"kubernetes.io/projected/8f46289b-9b6d-41df-aa0e-ad2b40216d0f-kube-api-access-bpw6d\") pod \"8f46289b-9b6d-41df-aa0e-ad2b40216d0f\" (UID: \"8f46289b-9b6d-41df-aa0e-ad2b40216d0f\") " Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.669398 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f46289b-9b6d-41df-aa0e-ad2b40216d0f-kube-api-access-bpw6d" (OuterVolumeSpecName: "kube-api-access-bpw6d") pod "8f46289b-9b6d-41df-aa0e-ad2b40216d0f" (UID: "8f46289b-9b6d-41df-aa0e-ad2b40216d0f"). InnerVolumeSpecName "kube-api-access-bpw6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:27:03 crc kubenswrapper[4767]: I0129 15:27:03.763635 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpw6d\" (UniqueName: \"kubernetes.io/projected/8f46289b-9b6d-41df-aa0e-ad2b40216d0f-kube-api-access-bpw6d\") on node \"crc\" DevicePath \"\"" Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.188874 4767 generic.go:334] "Generic (PLEG): container finished" podID="8f46289b-9b6d-41df-aa0e-ad2b40216d0f" containerID="357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec" exitCode=0 Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.188944 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fl7z6" Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.188972 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fl7z6" event={"ID":"8f46289b-9b6d-41df-aa0e-ad2b40216d0f","Type":"ContainerDied","Data":"357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec"} Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.189948 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fl7z6" event={"ID":"8f46289b-9b6d-41df-aa0e-ad2b40216d0f","Type":"ContainerDied","Data":"345fbb32e484196941eacd475c47becc9955589c4d1dd92f75020bb2b0ac792f"} Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.189978 4767 scope.go:117] "RemoveContainer" containerID="357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec" Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.190919 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tsfnc" event={"ID":"12bdee23-d478-434a-95bc-7d2c042f6906","Type":"ContainerStarted","Data":"562558214df47db3987118fdb06179ed1a3fed62ded2b1ac3c2fb403078b477f"} Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.190962 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tsfnc" event={"ID":"12bdee23-d478-434a-95bc-7d2c042f6906","Type":"ContainerStarted","Data":"7cf19913ce6fc1768714b57bab3f8cbc9c5403cdf925dfb632bde38c31f58ef9"} Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.204910 4767 scope.go:117] "RemoveContainer" containerID="357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec" Jan 29 15:27:04 crc kubenswrapper[4767]: E0129 15:27:04.205445 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec\": container with ID starting with 357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec not found: ID does not exist" containerID="357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec" Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.205478 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec"} err="failed to get container status \"357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec\": rpc error: code = NotFound desc = could not find container \"357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec\": container with ID starting with 357e7841b76c1003b5e5fa97145ba23cdc561056ba1b66876f80818544ebc4ec not found: ID does not exist" Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.209590 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tsfnc" podStartSLOduration=2.145064974 podStartE2EDuration="2.209571161s" podCreationTimestamp="2026-01-29 15:27:02 +0000 UTC" firstStartedPulling="2026-01-29 15:27:03.520148497 +0000 UTC m=+1579.330465536" lastFinishedPulling="2026-01-29 15:27:03.584654684 +0000 UTC m=+1579.394971723" observedRunningTime="2026-01-29 15:27:04.206647367 +0000 UTC m=+1580.016964406" watchObservedRunningTime="2026-01-29 15:27:04.209571161 +0000 UTC m=+1580.019888200" Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.233607 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fl7z6"] Jan 29 15:27:04 crc kubenswrapper[4767]: I0129 15:27:04.239013 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-fl7z6"] Jan 29 15:27:05 crc kubenswrapper[4767]: I0129 15:27:05.025691 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f46289b-9b6d-41df-aa0e-ad2b40216d0f" path="/var/lib/kubelet/pods/8f46289b-9b6d-41df-aa0e-ad2b40216d0f/volumes" Jan 29 15:27:13 crc kubenswrapper[4767]: I0129 15:27:13.096423 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:13 crc kubenswrapper[4767]: I0129 15:27:13.096808 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:13 crc kubenswrapper[4767]: I0129 15:27:13.134854 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:13 crc kubenswrapper[4767]: I0129 15:27:13.287405 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tsfnc" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.293034 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm"] Jan 29 15:27:19 crc kubenswrapper[4767]: E0129 15:27:19.293815 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f46289b-9b6d-41df-aa0e-ad2b40216d0f" containerName="registry-server" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.293830 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f46289b-9b6d-41df-aa0e-ad2b40216d0f" containerName="registry-server" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.294008 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f46289b-9b6d-41df-aa0e-ad2b40216d0f" containerName="registry-server" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.294946 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.296948 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8hxwg" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.304223 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm"] Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.370400 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-util\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.370473 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-bundle\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.370535 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d98xx\" (UniqueName: \"kubernetes.io/projected/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-kube-api-access-d98xx\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.471507 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-bundle\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.471615 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d98xx\" (UniqueName: \"kubernetes.io/projected/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-kube-api-access-d98xx\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.471662 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-util\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.472106 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-bundle\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.472124 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-util\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.493532 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d98xx\" (UniqueName: \"kubernetes.io/projected/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-kube-api-access-d98xx\") pod \"278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:19 crc kubenswrapper[4767]: I0129 15:27:19.616659 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:20 crc kubenswrapper[4767]: I0129 15:27:20.104354 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm"] Jan 29 15:27:20 crc kubenswrapper[4767]: I0129 15:27:20.299465 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" event={"ID":"ee2c1e3b-70b5-4479-a3f2-aa33aba08243","Type":"ContainerStarted","Data":"349cee132869c0ae01310675a6a11b066d3149867ab702f92110b2d960838b45"} Jan 29 15:27:20 crc kubenswrapper[4767]: I0129 15:27:20.299513 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" event={"ID":"ee2c1e3b-70b5-4479-a3f2-aa33aba08243","Type":"ContainerStarted","Data":"25043bcdb420d678c105e21eb43377da981eab74f2c88ba22bc8fa91401d7095"} Jan 29 15:27:21 crc kubenswrapper[4767]: I0129 15:27:21.306736 4767 generic.go:334] "Generic (PLEG): container finished" podID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerID="349cee132869c0ae01310675a6a11b066d3149867ab702f92110b2d960838b45" exitCode=0 Jan 29 15:27:21 crc kubenswrapper[4767]: I0129 15:27:21.307060 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" event={"ID":"ee2c1e3b-70b5-4479-a3f2-aa33aba08243","Type":"ContainerDied","Data":"349cee132869c0ae01310675a6a11b066d3149867ab702f92110b2d960838b45"} Jan 29 15:27:22 crc kubenswrapper[4767]: I0129 15:27:22.321767 4767 generic.go:334] "Generic (PLEG): container finished" podID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerID="6b78fb9c8473d50d20671dc52a0cd3c04bbedf91523d49d280a863e994dc5d26" exitCode=0 Jan 29 15:27:22 crc kubenswrapper[4767]: I0129 15:27:22.321878 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" event={"ID":"ee2c1e3b-70b5-4479-a3f2-aa33aba08243","Type":"ContainerDied","Data":"6b78fb9c8473d50d20671dc52a0cd3c04bbedf91523d49d280a863e994dc5d26"} Jan 29 15:27:23 crc kubenswrapper[4767]: I0129 15:27:23.334343 4767 generic.go:334] "Generic (PLEG): container finished" podID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerID="202750e17e812d1544edaea7c8b021c21fb945d96cd3a9ec6fe92ea50a3abb7e" exitCode=0 Jan 29 15:27:23 crc kubenswrapper[4767]: I0129 15:27:23.334406 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" event={"ID":"ee2c1e3b-70b5-4479-a3f2-aa33aba08243","Type":"ContainerDied","Data":"202750e17e812d1544edaea7c8b021c21fb945d96cd3a9ec6fe92ea50a3abb7e"} Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.624606 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.743493 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-bundle\") pod \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.743583 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d98xx\" (UniqueName: \"kubernetes.io/projected/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-kube-api-access-d98xx\") pod \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.743699 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-util\") pod \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\" (UID: \"ee2c1e3b-70b5-4479-a3f2-aa33aba08243\") " Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.744283 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-bundle" (OuterVolumeSpecName: "bundle") pod "ee2c1e3b-70b5-4479-a3f2-aa33aba08243" (UID: "ee2c1e3b-70b5-4479-a3f2-aa33aba08243"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.750144 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-kube-api-access-d98xx" (OuterVolumeSpecName: "kube-api-access-d98xx") pod "ee2c1e3b-70b5-4479-a3f2-aa33aba08243" (UID: "ee2c1e3b-70b5-4479-a3f2-aa33aba08243"). InnerVolumeSpecName "kube-api-access-d98xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.757590 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-util" (OuterVolumeSpecName: "util") pod "ee2c1e3b-70b5-4479-a3f2-aa33aba08243" (UID: "ee2c1e3b-70b5-4479-a3f2-aa33aba08243"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.844496 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d98xx\" (UniqueName: \"kubernetes.io/projected/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-kube-api-access-d98xx\") on node \"crc\" DevicePath \"\"" Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.844565 4767 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-util\") on node \"crc\" DevicePath \"\"" Jan 29 15:27:24 crc kubenswrapper[4767]: I0129 15:27:24.844577 4767 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ee2c1e3b-70b5-4479-a3f2-aa33aba08243-bundle\") on node \"crc\" DevicePath \"\"" Jan 29 15:27:25 crc kubenswrapper[4767]: I0129 15:27:25.352580 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" event={"ID":"ee2c1e3b-70b5-4479-a3f2-aa33aba08243","Type":"ContainerDied","Data":"25043bcdb420d678c105e21eb43377da981eab74f2c88ba22bc8fa91401d7095"} Jan 29 15:27:25 crc kubenswrapper[4767]: I0129 15:27:25.352631 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25043bcdb420d678c105e21eb43377da981eab74f2c88ba22bc8fa91401d7095" Jan 29 15:27:25 crc kubenswrapper[4767]: I0129 15:27:25.352680 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.423951 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847"] Jan 29 15:27:31 crc kubenswrapper[4767]: E0129 15:27:31.424712 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerName="util" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.424726 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerName="util" Jan 29 15:27:31 crc kubenswrapper[4767]: E0129 15:27:31.424740 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerName="extract" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.424746 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerName="extract" Jan 29 15:27:31 crc kubenswrapper[4767]: E0129 15:27:31.424754 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerName="pull" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.424760 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerName="pull" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.424915 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2c1e3b-70b5-4479-a3f2-aa33aba08243" containerName="extract" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.425392 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.429451 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-dx6pz" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.442167 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg5j4\" (UniqueName: \"kubernetes.io/projected/bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e-kube-api-access-zg5j4\") pod \"openstack-operator-controller-init-68476f6fbd-2l847\" (UID: \"bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e\") " pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.449673 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847"] Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.543725 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg5j4\" (UniqueName: \"kubernetes.io/projected/bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e-kube-api-access-zg5j4\") pod \"openstack-operator-controller-init-68476f6fbd-2l847\" (UID: \"bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e\") " pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.563730 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg5j4\" (UniqueName: \"kubernetes.io/projected/bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e-kube-api-access-zg5j4\") pod \"openstack-operator-controller-init-68476f6fbd-2l847\" (UID: \"bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e\") " pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" Jan 29 15:27:31 crc kubenswrapper[4767]: I0129 15:27:31.742989 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" Jan 29 15:27:32 crc kubenswrapper[4767]: I0129 15:27:32.172949 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847"] Jan 29 15:27:32 crc kubenswrapper[4767]: I0129 15:27:32.399108 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" event={"ID":"bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e","Type":"ContainerStarted","Data":"483788b8c1e9623f0c693b5448e4cbe0111d5877225e14b30428add5cd0e247a"} Jan 29 15:27:36 crc kubenswrapper[4767]: I0129 15:27:36.433192 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" event={"ID":"bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e","Type":"ContainerStarted","Data":"cf4d4b8657451b64ed475e1cdac330a761615fde31b59e04294b0729f55faccf"} Jan 29 15:27:36 crc kubenswrapper[4767]: I0129 15:27:36.433668 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" Jan 29 15:27:36 crc kubenswrapper[4767]: I0129 15:27:36.470560 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" podStartSLOduration=1.72777376 podStartE2EDuration="5.470513435s" podCreationTimestamp="2026-01-29 15:27:31 +0000 UTC" firstStartedPulling="2026-01-29 15:27:32.179675325 +0000 UTC m=+1607.989992364" lastFinishedPulling="2026-01-29 15:27:35.922415 +0000 UTC m=+1611.732732039" observedRunningTime="2026-01-29 15:27:36.465934674 +0000 UTC m=+1612.276251713" watchObservedRunningTime="2026-01-29 15:27:36.470513435 +0000 UTC m=+1612.280830464" Jan 29 15:27:41 crc kubenswrapper[4767]: I0129 15:27:41.746600 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-68476f6fbd-2l847" Jan 29 15:27:42 crc kubenswrapper[4767]: I0129 15:27:42.805268 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:27:42 crc kubenswrapper[4767]: I0129 15:27:42.805612 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:27:52 crc kubenswrapper[4767]: I0129 15:27:52.856150 4767 patch_prober.go:28] interesting pod/controller-manager-66b4b547f8-ng4bb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.74:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 15:27:52 crc kubenswrapper[4767]: I0129 15:27:52.856684 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" podUID="b5e054e7-e150-4190-910b-5f95da1c9155" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.74:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 15:27:52 crc kubenswrapper[4767]: I0129 15:27:52.856217 4767 patch_prober.go:28] interesting pod/controller-manager-66b4b547f8-ng4bb container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.74:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 29 15:27:52 crc kubenswrapper[4767]: I0129 15:27:52.857083 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-66b4b547f8-ng4bb" podUID="b5e054e7-e150-4190-910b-5f95da1c9155" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.74:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.344508 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.345762 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.347832 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6j4ts" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.365718 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.376664 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.377469 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.380863 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-bzsfp" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.384336 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.398472 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.399464 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.402013 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hlsp2" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.411923 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.412838 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.421053 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.421709 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8bvz\" (UniqueName: \"kubernetes.io/projected/4be7d2d9-0fa7-4783-987a-7dd06823b118-kube-api-access-t8bvz\") pod \"cinder-operator-controller-manager-858d89fd-v5bpj\" (UID: \"4be7d2d9-0fa7-4783-987a-7dd06823b118\") " pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.421775 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-z294k" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.421818 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlwb7\" (UniqueName: \"kubernetes.io/projected/157b230c-4bbc-4a62-841e-60acba588f1e-kube-api-access-nlwb7\") pod \"barbican-operator-controller-manager-7d6fdb96dc-gkq4d\" (UID: \"157b230c-4bbc-4a62-841e-60acba588f1e\") " pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.429881 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.430558 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.438671 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-mntgm" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.449600 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.469802 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.492933 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.495263 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.497428 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-xkz2q" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.508446 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.522642 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-br5bs"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.523749 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfl5j\" (UniqueName: \"kubernetes.io/projected/a1b7dc5f-dccb-44e8-b390-daa6f9e773b3-kube-api-access-dfl5j\") pod \"heat-operator-controller-manager-d8b84fbc-9zkzh\" (UID: \"a1b7dc5f-dccb-44e8-b390-daa6f9e773b3\") " pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.523792 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlwb7\" (UniqueName: \"kubernetes.io/projected/157b230c-4bbc-4a62-841e-60acba588f1e-kube-api-access-nlwb7\") pod \"barbican-operator-controller-manager-7d6fdb96dc-gkq4d\" (UID: \"157b230c-4bbc-4a62-841e-60acba588f1e\") " pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.523814 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsg64\" (UniqueName: \"kubernetes.io/projected/49292974-81e4-4eba-8dc8-d6ac21111b9a-kube-api-access-dsg64\") pod \"designate-operator-controller-manager-dd77988f8-bvxsd\" (UID: \"49292974-81e4-4eba-8dc8-d6ac21111b9a\") " pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.523859 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8bvz\" (UniqueName: \"kubernetes.io/projected/4be7d2d9-0fa7-4783-987a-7dd06823b118-kube-api-access-t8bvz\") pod \"cinder-operator-controller-manager-858d89fd-v5bpj\" (UID: \"4be7d2d9-0fa7-4783-987a-7dd06823b118\") " pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.523887 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94fkc\" (UniqueName: \"kubernetes.io/projected/440d7f0e-a462-453a-b41f-ccdb1521bb20-kube-api-access-94fkc\") pod \"glance-operator-controller-manager-f8c4db9df-r4nvx\" (UID: \"440d7f0e-a462-453a-b41f-ccdb1521bb20\") " pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.524366 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.526549 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-d296w" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.527386 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.527510 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-br5bs"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.544083 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.544818 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.549329 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-j29n6" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.559609 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.559849 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlwb7\" (UniqueName: \"kubernetes.io/projected/157b230c-4bbc-4a62-841e-60acba588f1e-kube-api-access-nlwb7\") pod \"barbican-operator-controller-manager-7d6fdb96dc-gkq4d\" (UID: \"157b230c-4bbc-4a62-841e-60acba588f1e\") " pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.566951 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.567884 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.576965 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.578211 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8vp8d" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.578689 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.585046 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.585825 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.589302 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-fbkr2" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.589521 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-mg8gh" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.596130 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.598954 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8bvz\" (UniqueName: \"kubernetes.io/projected/4be7d2d9-0fa7-4783-987a-7dd06823b118-kube-api-access-t8bvz\") pod \"cinder-operator-controller-manager-858d89fd-v5bpj\" (UID: \"4be7d2d9-0fa7-4783-987a-7dd06823b118\") " pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.612005 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.626799 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94fkc\" (UniqueName: \"kubernetes.io/projected/440d7f0e-a462-453a-b41f-ccdb1521bb20-kube-api-access-94fkc\") pod \"glance-operator-controller-manager-f8c4db9df-r4nvx\" (UID: \"440d7f0e-a462-453a-b41f-ccdb1521bb20\") " pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.626855 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvrvw\" (UniqueName: \"kubernetes.io/projected/6e272311-00bb-4cfb-92b5-2cf77eb483f8-kube-api-access-rvrvw\") pod \"keystone-operator-controller-manager-8ccc8547b-gd2cz\" (UID: \"6e272311-00bb-4cfb-92b5-2cf77eb483f8\") " pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.626883 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.626926 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k55cg\" (UniqueName: \"kubernetes.io/projected/97c4098d-defd-4156-b48b-e26dd157f709-kube-api-access-k55cg\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.626952 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcmb8\" (UniqueName: \"kubernetes.io/projected/3350030b-3cea-4a51-9fdd-34408ad506eb-kube-api-access-wcmb8\") pod \"mariadb-operator-controller-manager-67bf948998-rlgl7\" (UID: \"3350030b-3cea-4a51-9fdd-34408ad506eb\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.626973 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dt4f\" (UniqueName: \"kubernetes.io/projected/a53807b1-819e-4e75-9f4b-ebda7fc7b415-kube-api-access-4dt4f\") pod \"horizon-operator-controller-manager-5fb775575f-6gtvq\" (UID: \"a53807b1-819e-4e75-9f4b-ebda7fc7b415\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.626997 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd678\" (UniqueName: \"kubernetes.io/projected/68ca9f6a-ab9f-4188-a45c-23e6c1303a20-kube-api-access-qd678\") pod \"ironic-operator-controller-manager-866c9d5b98-hvtlx\" (UID: \"68ca9f6a-ab9f-4188-a45c-23e6c1303a20\") " pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.627042 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfl5j\" (UniqueName: \"kubernetes.io/projected/a1b7dc5f-dccb-44e8-b390-daa6f9e773b3-kube-api-access-dfl5j\") pod \"heat-operator-controller-manager-d8b84fbc-9zkzh\" (UID: \"a1b7dc5f-dccb-44e8-b390-daa6f9e773b3\") " pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.627071 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsg64\" (UniqueName: \"kubernetes.io/projected/49292974-81e4-4eba-8dc8-d6ac21111b9a-kube-api-access-dsg64\") pod \"designate-operator-controller-manager-dd77988f8-bvxsd\" (UID: \"49292974-81e4-4eba-8dc8-d6ac21111b9a\") " pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.627128 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgcn7\" (UniqueName: \"kubernetes.io/projected/eba759c2-d8c7-4956-a47c-def368a54bf1-kube-api-access-lgcn7\") pod \"manila-operator-controller-manager-76c896469f-xxrwp\" (UID: \"eba759c2-d8c7-4956-a47c-def368a54bf1\") " pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.632270 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.638280 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.639300 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.644925 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.652815 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bvjfx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.666956 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.694114 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.694690 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfl5j\" (UniqueName: \"kubernetes.io/projected/a1b7dc5f-dccb-44e8-b390-daa6f9e773b3-kube-api-access-dfl5j\") pod \"heat-operator-controller-manager-d8b84fbc-9zkzh\" (UID: \"a1b7dc5f-dccb-44e8-b390-daa6f9e773b3\") " pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.695253 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.706654 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsg64\" (UniqueName: \"kubernetes.io/projected/49292974-81e4-4eba-8dc8-d6ac21111b9a-kube-api-access-dsg64\") pod \"designate-operator-controller-manager-dd77988f8-bvxsd\" (UID: \"49292974-81e4-4eba-8dc8-d6ac21111b9a\") " pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.716522 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94fkc\" (UniqueName: \"kubernetes.io/projected/440d7f0e-a462-453a-b41f-ccdb1521bb20-kube-api-access-94fkc\") pod \"glance-operator-controller-manager-f8c4db9df-r4nvx\" (UID: \"440d7f0e-a462-453a-b41f-ccdb1521bb20\") " pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.717620 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-s67m4" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728348 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k55cg\" (UniqueName: \"kubernetes.io/projected/97c4098d-defd-4156-b48b-e26dd157f709-kube-api-access-k55cg\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728399 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcmb8\" (UniqueName: \"kubernetes.io/projected/3350030b-3cea-4a51-9fdd-34408ad506eb-kube-api-access-wcmb8\") pod \"mariadb-operator-controller-manager-67bf948998-rlgl7\" (UID: \"3350030b-3cea-4a51-9fdd-34408ad506eb\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728420 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dt4f\" (UniqueName: \"kubernetes.io/projected/a53807b1-819e-4e75-9f4b-ebda7fc7b415-kube-api-access-4dt4f\") pod \"horizon-operator-controller-manager-5fb775575f-6gtvq\" (UID: \"a53807b1-819e-4e75-9f4b-ebda7fc7b415\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728439 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd678\" (UniqueName: \"kubernetes.io/projected/68ca9f6a-ab9f-4188-a45c-23e6c1303a20-kube-api-access-qd678\") pod \"ironic-operator-controller-manager-866c9d5b98-hvtlx\" (UID: \"68ca9f6a-ab9f-4188-a45c-23e6c1303a20\") " pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728501 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmc22\" (UniqueName: \"kubernetes.io/projected/a7f3205f-40e9-4eed-8305-4e0ca0ebe75c-kube-api-access-bmc22\") pod \"neutron-operator-controller-manager-7c7cc6ff45-zwfl4\" (UID: \"a7f3205f-40e9-4eed-8305-4e0ca0ebe75c\") " pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728530 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7972v\" (UniqueName: \"kubernetes.io/projected/833297bf-cdda-441d-8e03-de62a3b0c5b3-kube-api-access-7972v\") pod \"nova-operator-controller-manager-68cb478976-gbtpt\" (UID: \"833297bf-cdda-441d-8e03-de62a3b0c5b3\") " pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728553 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgcn7\" (UniqueName: \"kubernetes.io/projected/eba759c2-d8c7-4956-a47c-def368a54bf1-kube-api-access-lgcn7\") pod \"manila-operator-controller-manager-76c896469f-xxrwp\" (UID: \"eba759c2-d8c7-4956-a47c-def368a54bf1\") " pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728594 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvrvw\" (UniqueName: \"kubernetes.io/projected/6e272311-00bb-4cfb-92b5-2cf77eb483f8-kube-api-access-rvrvw\") pod \"keystone-operator-controller-manager-8ccc8547b-gd2cz\" (UID: \"6e272311-00bb-4cfb-92b5-2cf77eb483f8\") " pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.728613 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:27:59 crc kubenswrapper[4767]: E0129 15:27:59.729604 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 15:27:59 crc kubenswrapper[4767]: E0129 15:27:59.729694 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert podName:97c4098d-defd-4156-b48b-e26dd157f709 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:00.22966684 +0000 UTC m=+1636.039983879 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert") pod "infra-operator-controller-manager-79955696d6-br5bs" (UID: "97c4098d-defd-4156-b48b-e26dd157f709") : secret "infra-operator-webhook-server-cert" not found Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.737756 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.753101 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.765228 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.768609 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k55cg\" (UniqueName: \"kubernetes.io/projected/97c4098d-defd-4156-b48b-e26dd157f709-kube-api-access-k55cg\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.770121 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.776365 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.783689 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dt4f\" (UniqueName: \"kubernetes.io/projected/a53807b1-819e-4e75-9f4b-ebda7fc7b415-kube-api-access-4dt4f\") pod \"horizon-operator-controller-manager-5fb775575f-6gtvq\" (UID: \"a53807b1-819e-4e75-9f4b-ebda7fc7b415\") " pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.784148 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.806575 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgcn7\" (UniqueName: \"kubernetes.io/projected/eba759c2-d8c7-4956-a47c-def368a54bf1-kube-api-access-lgcn7\") pod \"manila-operator-controller-manager-76c896469f-xxrwp\" (UID: \"eba759c2-d8c7-4956-a47c-def368a54bf1\") " pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.806636 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvrvw\" (UniqueName: \"kubernetes.io/projected/6e272311-00bb-4cfb-92b5-2cf77eb483f8-kube-api-access-rvrvw\") pod \"keystone-operator-controller-manager-8ccc8547b-gd2cz\" (UID: \"6e272311-00bb-4cfb-92b5-2cf77eb483f8\") " pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.807115 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcmb8\" (UniqueName: \"kubernetes.io/projected/3350030b-3cea-4a51-9fdd-34408ad506eb-kube-api-access-wcmb8\") pod \"mariadb-operator-controller-manager-67bf948998-rlgl7\" (UID: \"3350030b-3cea-4a51-9fdd-34408ad506eb\") " pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.808002 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-9l96t" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.818123 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.818509 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.819382 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.826211 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd678\" (UniqueName: \"kubernetes.io/projected/68ca9f6a-ab9f-4188-a45c-23e6c1303a20-kube-api-access-qd678\") pod \"ironic-operator-controller-manager-866c9d5b98-hvtlx\" (UID: \"68ca9f6a-ab9f-4188-a45c-23e6c1303a20\") " pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.833548 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86wss\" (UniqueName: \"kubernetes.io/projected/ca1f1109-544f-490f-9bf5-9d05241d30f2-kube-api-access-86wss\") pod \"octavia-operator-controller-manager-68f8cb846c-cs4g8\" (UID: \"ca1f1109-544f-490f-9bf5-9d05241d30f2\") " pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.833622 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmc22\" (UniqueName: \"kubernetes.io/projected/a7f3205f-40e9-4eed-8305-4e0ca0ebe75c-kube-api-access-bmc22\") pod \"neutron-operator-controller-manager-7c7cc6ff45-zwfl4\" (UID: \"a7f3205f-40e9-4eed-8305-4e0ca0ebe75c\") " pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.833648 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7972v\" (UniqueName: \"kubernetes.io/projected/833297bf-cdda-441d-8e03-de62a3b0c5b3-kube-api-access-7972v\") pod \"nova-operator-controller-manager-68cb478976-gbtpt\" (UID: \"833297bf-cdda-441d-8e03-de62a3b0c5b3\") " pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.834369 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.861516 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.862563 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.863308 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmc22\" (UniqueName: \"kubernetes.io/projected/a7f3205f-40e9-4eed-8305-4e0ca0ebe75c-kube-api-access-bmc22\") pod \"neutron-operator-controller-manager-7c7cc6ff45-zwfl4\" (UID: \"a7f3205f-40e9-4eed-8305-4e0ca0ebe75c\") " pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.865599 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7972v\" (UniqueName: \"kubernetes.io/projected/833297bf-cdda-441d-8e03-de62a3b0c5b3-kube-api-access-7972v\") pod \"nova-operator-controller-manager-68cb478976-gbtpt\" (UID: \"833297bf-cdda-441d-8e03-de62a3b0c5b3\") " pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.866254 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-w8hqx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.902612 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.931448 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.934325 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.936382 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7flcc\" (UniqueName: \"kubernetes.io/projected/0d956603-d824-432d-b1db-6574d2bde092-kube-api-access-7flcc\") pod \"ovn-operator-controller-manager-788c46999f-gbkxz\" (UID: \"0d956603-d824-432d-b1db-6574d2bde092\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.936509 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86wss\" (UniqueName: \"kubernetes.io/projected/ca1f1109-544f-490f-9bf5-9d05241d30f2-kube-api-access-86wss\") pod \"octavia-operator-controller-manager-68f8cb846c-cs4g8\" (UID: \"ca1f1109-544f-490f-9bf5-9d05241d30f2\") " pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.938759 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.939920 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.942474 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.945779 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-hw56q" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.957351 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.962149 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.966383 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86wss\" (UniqueName: \"kubernetes.io/projected/ca1f1109-544f-490f-9bf5-9d05241d30f2-kube-api-access-86wss\") pod \"octavia-operator-controller-manager-68f8cb846c-cs4g8\" (UID: \"ca1f1109-544f-490f-9bf5-9d05241d30f2\") " pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.975782 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh"] Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.976639 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.980292 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-d9bn7" Jan 29 15:27:59 crc kubenswrapper[4767]: I0129 15:27:59.989443 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.018733 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.022526 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.030935 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.037420 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.037493 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ktkr\" (UniqueName: \"kubernetes.io/projected/adc494cd-a7d7-4104-9d5b-165115dfe504-kube-api-access-9ktkr\") pod \"swift-operator-controller-manager-6f7455757b-pc4p9\" (UID: \"adc494cd-a7d7-4104-9d5b-165115dfe504\") " pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.037523 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7flcc\" (UniqueName: \"kubernetes.io/projected/0d956603-d824-432d-b1db-6574d2bde092-kube-api-access-7flcc\") pod \"ovn-operator-controller-manager-788c46999f-gbkxz\" (UID: \"0d956603-d824-432d-b1db-6574d2bde092\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.037611 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sdmf\" (UniqueName: \"kubernetes.io/projected/1a51e166-62bf-4181-b44f-62d52b543f2e-kube-api-access-7sdmf\") pod \"placement-operator-controller-manager-5b964cf4cd-fsgxh\" (UID: \"1a51e166-62bf-4181-b44f-62d52b543f2e\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.037643 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szx4f\" (UniqueName: \"kubernetes.io/projected/dc435770-6862-4056-898e-ee9fbf4850d3-kube-api-access-szx4f\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.040134 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-k7h8q" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.063433 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7flcc\" (UniqueName: \"kubernetes.io/projected/0d956603-d824-432d-b1db-6574d2bde092-kube-api-access-7flcc\") pod \"ovn-operator-controller-manager-788c46999f-gbkxz\" (UID: \"0d956603-d824-432d-b1db-6574d2bde092\") " pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.068145 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.089904 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.115404 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.116271 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.119636 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-rs9cw" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.121467 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.128962 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.129778 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.136724 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bvtxb" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.138865 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ktkr\" (UniqueName: \"kubernetes.io/projected/adc494cd-a7d7-4104-9d5b-165115dfe504-kube-api-access-9ktkr\") pod \"swift-operator-controller-manager-6f7455757b-pc4p9\" (UID: \"adc494cd-a7d7-4104-9d5b-165115dfe504\") " pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.138927 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fnmt\" (UniqueName: \"kubernetes.io/projected/b186845a-5b3b-42ce-aad0-0269580f5b07-kube-api-access-8fnmt\") pod \"telemetry-operator-controller-manager-6cf8c44c7-nzz8p\" (UID: \"b186845a-5b3b-42ce-aad0-0269580f5b07\") " pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.138971 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fshcw\" (UniqueName: \"kubernetes.io/projected/7b712b0e-3913-4421-a6c8-90991284909a-kube-api-access-fshcw\") pod \"test-operator-controller-manager-56f8bfcd9f-dbphp\" (UID: \"7b712b0e-3913-4421-a6c8-90991284909a\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.138994 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sdmf\" (UniqueName: \"kubernetes.io/projected/1a51e166-62bf-4181-b44f-62d52b543f2e-kube-api-access-7sdmf\") pod \"placement-operator-controller-manager-5b964cf4cd-fsgxh\" (UID: \"1a51e166-62bf-4181-b44f-62d52b543f2e\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.139020 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szx4f\" (UniqueName: \"kubernetes.io/projected/dc435770-6862-4056-898e-ee9fbf4850d3-kube-api-access-szx4f\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.139044 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.139147 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.139184 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert podName:dc435770-6862-4056-898e-ee9fbf4850d3 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:00.639171375 +0000 UTC m=+1636.449488414 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" (UID: "dc435770-6862-4056-898e-ee9fbf4850d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.148261 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.179974 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.180826 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.185211 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-5zwwg" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.185726 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.201483 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szx4f\" (UniqueName: \"kubernetes.io/projected/dc435770-6862-4056-898e-ee9fbf4850d3-kube-api-access-szx4f\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.202122 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ktkr\" (UniqueName: \"kubernetes.io/projected/adc494cd-a7d7-4104-9d5b-165115dfe504-kube-api-access-9ktkr\") pod \"swift-operator-controller-manager-6f7455757b-pc4p9\" (UID: \"adc494cd-a7d7-4104-9d5b-165115dfe504\") " pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.206023 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sdmf\" (UniqueName: \"kubernetes.io/projected/1a51e166-62bf-4181-b44f-62d52b543f2e-kube-api-access-7sdmf\") pod \"placement-operator-controller-manager-5b964cf4cd-fsgxh\" (UID: \"1a51e166-62bf-4181-b44f-62d52b543f2e\") " pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.245877 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fshcw\" (UniqueName: \"kubernetes.io/projected/7b712b0e-3913-4421-a6c8-90991284909a-kube-api-access-fshcw\") pod \"test-operator-controller-manager-56f8bfcd9f-dbphp\" (UID: \"7b712b0e-3913-4421-a6c8-90991284909a\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.245960 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsf78\" (UniqueName: \"kubernetes.io/projected/e63115ec-b52a-4e9f-89b5-76004e0f6316-kube-api-access-bsf78\") pod \"watcher-operator-controller-manager-59f4c7d7c4-g6hsv\" (UID: \"e63115ec-b52a-4e9f-89b5-76004e0f6316\") " pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.246012 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.246049 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fnmt\" (UniqueName: \"kubernetes.io/projected/b186845a-5b3b-42ce-aad0-0269580f5b07-kube-api-access-8fnmt\") pod \"telemetry-operator-controller-manager-6cf8c44c7-nzz8p\" (UID: \"b186845a-5b3b-42ce-aad0-0269580f5b07\") " pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.246726 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.246760 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert podName:97c4098d-defd-4156-b48b-e26dd157f709 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:01.246748276 +0000 UTC m=+1637.057065315 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert") pod "infra-operator-controller-manager-79955696d6-br5bs" (UID: "97c4098d-defd-4156-b48b-e26dd157f709") : secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.251096 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.254966 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.255996 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.259423 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.259637 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-m6g6h" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.259800 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.270261 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.275830 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.289244 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fnmt\" (UniqueName: \"kubernetes.io/projected/b186845a-5b3b-42ce-aad0-0269580f5b07-kube-api-access-8fnmt\") pod \"telemetry-operator-controller-manager-6cf8c44c7-nzz8p\" (UID: \"b186845a-5b3b-42ce-aad0-0269580f5b07\") " pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.294013 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fshcw\" (UniqueName: \"kubernetes.io/projected/7b712b0e-3913-4421-a6c8-90991284909a-kube-api-access-fshcw\") pod \"test-operator-controller-manager-56f8bfcd9f-dbphp\" (UID: \"7b712b0e-3913-4421-a6c8-90991284909a\") " pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.303622 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.304549 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.306689 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-hfxtr" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.320193 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.326343 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.348270 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk2j5\" (UniqueName: \"kubernetes.io/projected/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-kube-api-access-pk2j5\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.348665 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.348807 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsf78\" (UniqueName: \"kubernetes.io/projected/e63115ec-b52a-4e9f-89b5-76004e0f6316-kube-api-access-bsf78\") pod \"watcher-operator-controller-manager-59f4c7d7c4-g6hsv\" (UID: \"e63115ec-b52a-4e9f-89b5-76004e0f6316\") " pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.348853 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.348969 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfdbk\" (UniqueName: \"kubernetes.io/projected/f104967e-4b06-42b6-b9ea-c4edb1923e87-kube-api-access-dfdbk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xmjrr\" (UID: \"f104967e-4b06-42b6-b9ea-c4edb1923e87\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.368066 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsf78\" (UniqueName: \"kubernetes.io/projected/e63115ec-b52a-4e9f-89b5-76004e0f6316-kube-api-access-bsf78\") pod \"watcher-operator-controller-manager-59f4c7d7c4-g6hsv\" (UID: \"e63115ec-b52a-4e9f-89b5-76004e0f6316\") " pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.385679 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.456209 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk2j5\" (UniqueName: \"kubernetes.io/projected/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-kube-api-access-pk2j5\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.456325 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.456435 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.456480 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfdbk\" (UniqueName: \"kubernetes.io/projected/f104967e-4b06-42b6-b9ea-c4edb1923e87-kube-api-access-dfdbk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xmjrr\" (UID: \"f104967e-4b06-42b6-b9ea-c4edb1923e87\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.457112 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.457152 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:00.95713736 +0000 UTC m=+1636.767454399 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.457313 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.457335 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:00.957328915 +0000 UTC m=+1636.767645954 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "metrics-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.476748 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfdbk\" (UniqueName: \"kubernetes.io/projected/f104967e-4b06-42b6-b9ea-c4edb1923e87-kube-api-access-dfdbk\") pod \"rabbitmq-cluster-operator-manager-668c99d594-xmjrr\" (UID: \"f104967e-4b06-42b6-b9ea-c4edb1923e87\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.482556 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk2j5\" (UniqueName: \"kubernetes.io/projected/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-kube-api-access-pk2j5\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.506515 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.553544 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.596232 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.598310 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d"] Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.659807 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.660005 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.660051 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert podName:dc435770-6862-4056-898e-ee9fbf4850d3 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:01.660038 +0000 UTC m=+1637.470355039 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" (UID: "dc435770-6862-4056-898e-ee9fbf4850d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.720624 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.966077 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: I0129 15:28:00.966480 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.966666 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.966716 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:01.96670071 +0000 UTC m=+1637.777017749 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "metrics-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.967131 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 15:28:00 crc kubenswrapper[4767]: E0129 15:28:00.967156 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:01.967148943 +0000 UTC m=+1637.777465972 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "webhook-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.010005 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq"] Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.022273 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda53807b1_819e_4e75_9f4b_ebda7fc7b415.slice/crio-ecaa537e18ab245676d38ea3c21716431298956745cce99cfd105b95e4f36b0e WatchSource:0}: Error finding container ecaa537e18ab245676d38ea3c21716431298956745cce99cfd105b95e4f36b0e: Status 404 returned error can't find the container with id ecaa537e18ab245676d38ea3c21716431298956745cce99cfd105b95e4f36b0e Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.045489 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.059746 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.068252 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj"] Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.069942 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4be7d2d9_0fa7_4783_987a_7dd06823b118.slice/crio-3fdf721f534a0e701ac9fe139f0f4a58cec8dbd9d5642886a51ede120edc9466 WatchSource:0}: Error finding container 3fdf721f534a0e701ac9fe139f0f4a58cec8dbd9d5642886a51ede120edc9466: Status 404 returned error can't find the container with id 3fdf721f534a0e701ac9fe139f0f4a58cec8dbd9d5642886a51ede120edc9466 Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.075623 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49292974_81e4_4eba_8dc8_d6ac21111b9a.slice/crio-fc79a7a72a7f4e867e36d2caf3dc4bc2ca5b73309edd8f6ad34408a0b905ff20 WatchSource:0}: Error finding container fc79a7a72a7f4e867e36d2caf3dc4bc2ca5b73309edd8f6ad34408a0b905ff20: Status 404 returned error can't find the container with id fc79a7a72a7f4e867e36d2caf3dc4bc2ca5b73309edd8f6ad34408a0b905ff20 Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.090755 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.192020 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.197128 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp"] Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.198578 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68ca9f6a_ab9f_4188_a45c_23e6c1303a20.slice/crio-75cb1233a9720d4ab8a367ede49a23c28b2df64a7513cea8d82d913a698c80bf WatchSource:0}: Error finding container 75cb1233a9720d4ab8a367ede49a23c28b2df64a7513cea8d82d913a698c80bf: Status 404 returned error can't find the container with id 75cb1233a9720d4ab8a367ede49a23c28b2df64a7513cea8d82d913a698c80bf Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.200266 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca1f1109_544f_490f_9bf5_9d05241d30f2.slice/crio-84e8c037cdeede114f267b5a6583269f24b1253fb607d3bbe2828cb8dee459dc WatchSource:0}: Error finding container 84e8c037cdeede114f267b5a6583269f24b1253fb607d3bbe2828cb8dee459dc: Status 404 returned error can't find the container with id 84e8c037cdeede114f267b5a6583269f24b1253fb607d3bbe2828cb8dee459dc Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.201709 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.206744 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.211641 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt"] Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.212043 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/manila-operator@sha256:485c920b3385679f1df13ba46707c204b4212ea23621cbc75b44c062da20e495,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lgcn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-76c896469f-xxrwp_openstack-operators(eba759c2-d8c7-4956-a47c-def368a54bf1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.214054 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" podUID="eba759c2-d8c7-4956-a47c-def368a54bf1" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.216136 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.271472 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.271627 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.271683 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert podName:97c4098d-defd-4156-b48b-e26dd157f709 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:03.271667503 +0000 UTC m=+1639.081984542 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert") pod "infra-operator-controller-manager-79955696d6-br5bs" (UID: "97c4098d-defd-4156-b48b-e26dd157f709") : secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.371509 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv"] Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.383634 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a51e166_62bf_4181_b44f_62d52b543f2e.slice/crio-7b51b85a3fc08c10ddf31c1fad80768223cc61630053ce7f83f81c6d47a995dc WatchSource:0}: Error finding container 7b51b85a3fc08c10ddf31c1fad80768223cc61630053ce7f83f81c6d47a995dc: Status 404 returned error can't find the container with id 7b51b85a3fc08c10ddf31c1fad80768223cc61630053ce7f83f81c6d47a995dc Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.385195 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p"] Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.385608 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7sdmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5b964cf4cd-fsgxh_openstack-operators(1a51e166-62bf-4181-b44f-62d52b543f2e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.387204 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" podUID="1a51e166-62bf-4181-b44f-62d52b543f2e" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.388668 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/watcher-operator@sha256:d23c69ab5c7d6c649fe9e23db98eae9b9de8dce4f4901511b2b764dd366d7c2c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bsf78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-59f4c7d7c4-g6hsv_openstack-operators(e63115ec-b52a-4e9f-89b5-76004e0f6316): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.389810 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" podUID="e63115ec-b52a-4e9f-89b5-76004e0f6316" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.393801 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/telemetry-operator@sha256:ee0236c7a8c8383b0a633b6f6e5f31200462ba68a51c45362836014c08c0c976,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8fnmt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6cf8c44c7-nzz8p_openstack-operators(b186845a-5b3b-42ce-aad0-0269580f5b07): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.395019 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" podUID="b186845a-5b3b-42ce-aad0-0269580f5b07" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.396050 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:ea7b72b648a5bde2eebd804c2a5c1608d448a4892176c1b8d000c1eef4bb92b4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7flcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-788c46999f-gbkxz_openstack-operators(0d956603-d824-432d-b1db-6574d2bde092): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.396207 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dfdbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-xmjrr_openstack-operators(f104967e-4b06-42b6-b9ea-c4edb1923e87): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.397155 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" podUID="0d956603-d824-432d-b1db-6574d2bde092" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.397311 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" podUID="f104967e-4b06-42b6-b9ea-c4edb1923e87" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.397366 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.412304 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp"] Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.412581 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod440d7f0e_a462_453a_b41f_ccdb1521bb20.slice/crio-fccdbe675fde532bd5b938e7281aafd7900674e79ea0424dc0f1e0cb4bec7fc6 WatchSource:0}: Error finding container fccdbe675fde532bd5b938e7281aafd7900674e79ea0424dc0f1e0cb4bec7fc6: Status 404 returned error can't find the container with id fccdbe675fde532bd5b938e7281aafd7900674e79ea0424dc0f1e0cb4bec7fc6 Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.415672 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/lmiccini/glance-operator@sha256:ebb3f9f6e871da3fdfdefdf4040964abcdc5f4c7dac961a27c85a80f37866f00,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-94fkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-f8c4db9df-r4nvx_openstack-operators(440d7f0e-a462-453a-b41f-ccdb1521bb20): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.418579 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" podUID="440d7f0e-a462-453a-b41f-ccdb1521bb20" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.420490 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz"] Jan 29 15:28:01 crc kubenswrapper[4767]: W0129 15:28:01.427090 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadc494cd_a7d7_4104_9d5b_165115dfe504.slice/crio-179faf63ef9f8bafed8ed06609397f7c39362ba2b2aa8c5b7eb97a3854e74d76 WatchSource:0}: Error finding container 179faf63ef9f8bafed8ed06609397f7c39362ba2b2aa8c5b7eb97a3854e74d76: Status 404 returned error can't find the container with id 179faf63ef9f8bafed8ed06609397f7c39362ba2b2aa8c5b7eb97a3854e74d76 Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.427952 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.438254 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.445555 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9"] Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.621529 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" event={"ID":"49292974-81e4-4eba-8dc8-d6ac21111b9a","Type":"ContainerStarted","Data":"fc79a7a72a7f4e867e36d2caf3dc4bc2ca5b73309edd8f6ad34408a0b905ff20"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.622908 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" event={"ID":"f104967e-4b06-42b6-b9ea-c4edb1923e87","Type":"ContainerStarted","Data":"8fcdcfbd9529060d705d46009edb360fc54e2ebbabb3bb30bf5380243aa75048"} Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.623910 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" podUID="f104967e-4b06-42b6-b9ea-c4edb1923e87" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.625539 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" event={"ID":"e63115ec-b52a-4e9f-89b5-76004e0f6316","Type":"ContainerStarted","Data":"0f469d29593cc287c843ff00b7b3457a7731b188c43b60418fc6db3837d03a99"} Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.628320 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/watcher-operator@sha256:d23c69ab5c7d6c649fe9e23db98eae9b9de8dce4f4901511b2b764dd366d7c2c\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" podUID="e63115ec-b52a-4e9f-89b5-76004e0f6316" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.628700 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" event={"ID":"a1b7dc5f-dccb-44e8-b390-daa6f9e773b3","Type":"ContainerStarted","Data":"a18b32c9df057e3f11e2bc56563a93c3dd9337cf0db628e4bb51240c336229fa"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.629985 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" event={"ID":"440d7f0e-a462-453a-b41f-ccdb1521bb20","Type":"ContainerStarted","Data":"fccdbe675fde532bd5b938e7281aafd7900674e79ea0424dc0f1e0cb4bec7fc6"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.631384 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" event={"ID":"6e272311-00bb-4cfb-92b5-2cf77eb483f8","Type":"ContainerStarted","Data":"2a7239d6c9dd0999d7ea5342f264b7e6865c9ee1ff44aacce6dc327a2ccb02af"} Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.631492 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/glance-operator@sha256:ebb3f9f6e871da3fdfdefdf4040964abcdc5f4c7dac961a27c85a80f37866f00\\\"\"" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" podUID="440d7f0e-a462-453a-b41f-ccdb1521bb20" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.632604 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" event={"ID":"833297bf-cdda-441d-8e03-de62a3b0c5b3","Type":"ContainerStarted","Data":"fa02ae4a9dce81a49504ff07a0c33aa60878adadc25108a5bede6a85c2047090"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.633593 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" event={"ID":"a53807b1-819e-4e75-9f4b-ebda7fc7b415","Type":"ContainerStarted","Data":"ecaa537e18ab245676d38ea3c21716431298956745cce99cfd105b95e4f36b0e"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.634815 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" event={"ID":"0d956603-d824-432d-b1db-6574d2bde092","Type":"ContainerStarted","Data":"bf475a76dab2e15aecf5c782df1a9a47d2c0f38cc1c2aa21b68e65b696907e3d"} Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.636541 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:ea7b72b648a5bde2eebd804c2a5c1608d448a4892176c1b8d000c1eef4bb92b4\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" podUID="0d956603-d824-432d-b1db-6574d2bde092" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.637453 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" event={"ID":"a7f3205f-40e9-4eed-8305-4e0ca0ebe75c","Type":"ContainerStarted","Data":"900ce4218de14abf9c8db78b252744aee240174de07959489f204d05cd45bdd4"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.640388 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" event={"ID":"7b712b0e-3913-4421-a6c8-90991284909a","Type":"ContainerStarted","Data":"884a606aff42467a55cb44aa154e0c912d3f5b81b1f7aa0e2fb6a29f7bf9a620"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.642181 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" event={"ID":"adc494cd-a7d7-4104-9d5b-165115dfe504","Type":"ContainerStarted","Data":"179faf63ef9f8bafed8ed06609397f7c39362ba2b2aa8c5b7eb97a3854e74d76"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.643944 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" event={"ID":"3350030b-3cea-4a51-9fdd-34408ad506eb","Type":"ContainerStarted","Data":"bf6fddf71a8f0634f93ce6844164aeffd21a2ba22de9de080536d3eeb0842ded"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.645667 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" event={"ID":"ca1f1109-544f-490f-9bf5-9d05241d30f2","Type":"ContainerStarted","Data":"84e8c037cdeede114f267b5a6583269f24b1253fb607d3bbe2828cb8dee459dc"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.647438 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" event={"ID":"eba759c2-d8c7-4956-a47c-def368a54bf1","Type":"ContainerStarted","Data":"2e8f04804bece5c41d31bdae762539062db9f6823f5506b7b920391f45ebb36b"} Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.649472 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/manila-operator@sha256:485c920b3385679f1df13ba46707c204b4212ea23621cbc75b44c062da20e495\\\"\"" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" podUID="eba759c2-d8c7-4956-a47c-def368a54bf1" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.650324 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" event={"ID":"68ca9f6a-ab9f-4188-a45c-23e6c1303a20","Type":"ContainerStarted","Data":"75cb1233a9720d4ab8a367ede49a23c28b2df64a7513cea8d82d913a698c80bf"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.660167 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" event={"ID":"157b230c-4bbc-4a62-841e-60acba588f1e","Type":"ContainerStarted","Data":"804ca4be0341969149c9a2580a65860109d87bce09cdb454d8e4603833ebaa41"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.665539 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" event={"ID":"1a51e166-62bf-4181-b44f-62d52b543f2e","Type":"ContainerStarted","Data":"7b51b85a3fc08c10ddf31c1fad80768223cc61630053ce7f83f81c6d47a995dc"} Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.667831 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" podUID="1a51e166-62bf-4181-b44f-62d52b543f2e" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.668620 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" event={"ID":"b186845a-5b3b-42ce-aad0-0269580f5b07","Type":"ContainerStarted","Data":"41a9334e36f7121835bf23fe0d81e4e4a0475def763e1733b146e0381e6f4138"} Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.671449 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/telemetry-operator@sha256:ee0236c7a8c8383b0a633b6f6e5f31200462ba68a51c45362836014c08c0c976\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" podUID="b186845a-5b3b-42ce-aad0-0269580f5b07" Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.674047 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" event={"ID":"4be7d2d9-0fa7-4783-987a-7dd06823b118","Type":"ContainerStarted","Data":"3fdf721f534a0e701ac9fe139f0f4a58cec8dbd9d5642886a51ede120edc9466"} Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.681717 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.682994 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.683046 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert podName:dc435770-6862-4056-898e-ee9fbf4850d3 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:03.6830295 +0000 UTC m=+1639.493346549 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" (UID: "dc435770-6862-4056-898e-ee9fbf4850d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.998662 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.999074 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.999169 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:03.999147332 +0000 UTC m=+1639.809464481 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "webhook-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.999196 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: E0129 15:28:01.999248 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:03.999232634 +0000 UTC m=+1639.809549673 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "metrics-server-cert" not found Jan 29 15:28:01 crc kubenswrapper[4767]: I0129 15:28:01.999100 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:02 crc kubenswrapper[4767]: E0129 15:28:02.697867 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/manila-operator@sha256:485c920b3385679f1df13ba46707c204b4212ea23621cbc75b44c062da20e495\\\"\"" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" podUID="eba759c2-d8c7-4956-a47c-def368a54bf1" Jan 29 15:28:02 crc kubenswrapper[4767]: E0129 15:28:02.698104 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" podUID="f104967e-4b06-42b6-b9ea-c4edb1923e87" Jan 29 15:28:02 crc kubenswrapper[4767]: E0129 15:28:02.698122 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/watcher-operator@sha256:d23c69ab5c7d6c649fe9e23db98eae9b9de8dce4f4901511b2b764dd366d7c2c\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" podUID="e63115ec-b52a-4e9f-89b5-76004e0f6316" Jan 29 15:28:02 crc kubenswrapper[4767]: E0129 15:28:02.698150 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e0824d5d461ada59715eb3048ed9394c80abba09c45503f8f90ee3b34e525488\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" podUID="1a51e166-62bf-4181-b44f-62d52b543f2e" Jan 29 15:28:02 crc kubenswrapper[4767]: E0129 15:28:02.698179 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/telemetry-operator@sha256:ee0236c7a8c8383b0a633b6f6e5f31200462ba68a51c45362836014c08c0c976\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" podUID="b186845a-5b3b-42ce-aad0-0269580f5b07" Jan 29 15:28:02 crc kubenswrapper[4767]: E0129 15:28:02.698190 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:ea7b72b648a5bde2eebd804c2a5c1608d448a4892176c1b8d000c1eef4bb92b4\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" podUID="0d956603-d824-432d-b1db-6574d2bde092" Jan 29 15:28:02 crc kubenswrapper[4767]: E0129 15:28:02.704792 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/lmiccini/glance-operator@sha256:ebb3f9f6e871da3fdfdefdf4040964abcdc5f4c7dac961a27c85a80f37866f00\\\"\"" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" podUID="440d7f0e-a462-453a-b41f-ccdb1521bb20" Jan 29 15:28:03 crc kubenswrapper[4767]: I0129 15:28:03.322653 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:03 crc kubenswrapper[4767]: E0129 15:28:03.322795 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:03 crc kubenswrapper[4767]: E0129 15:28:03.322842 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert podName:97c4098d-defd-4156-b48b-e26dd157f709 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:07.322829994 +0000 UTC m=+1643.133147033 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert") pod "infra-operator-controller-manager-79955696d6-br5bs" (UID: "97c4098d-defd-4156-b48b-e26dd157f709") : secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:03 crc kubenswrapper[4767]: I0129 15:28:03.728115 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:03 crc kubenswrapper[4767]: E0129 15:28:03.728344 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:03 crc kubenswrapper[4767]: E0129 15:28:03.728388 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert podName:dc435770-6862-4056-898e-ee9fbf4850d3 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:07.728375536 +0000 UTC m=+1643.538692575 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" (UID: "dc435770-6862-4056-898e-ee9fbf4850d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:04 crc kubenswrapper[4767]: I0129 15:28:04.031901 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:04 crc kubenswrapper[4767]: E0129 15:28:04.032067 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 15:28:04 crc kubenswrapper[4767]: I0129 15:28:04.032238 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:04 crc kubenswrapper[4767]: E0129 15:28:04.032261 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:08.032237007 +0000 UTC m=+1643.842554046 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "metrics-server-cert" not found Jan 29 15:28:04 crc kubenswrapper[4767]: E0129 15:28:04.032346 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 15:28:04 crc kubenswrapper[4767]: E0129 15:28:04.032396 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:08.032382851 +0000 UTC m=+1643.842699890 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "webhook-server-cert" not found Jan 29 15:28:07 crc kubenswrapper[4767]: I0129 15:28:07.402190 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:07 crc kubenswrapper[4767]: E0129 15:28:07.402433 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:07 crc kubenswrapper[4767]: E0129 15:28:07.402479 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert podName:97c4098d-defd-4156-b48b-e26dd157f709 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:15.402464207 +0000 UTC m=+1651.212781246 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert") pod "infra-operator-controller-manager-79955696d6-br5bs" (UID: "97c4098d-defd-4156-b48b-e26dd157f709") : secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:07 crc kubenswrapper[4767]: I0129 15:28:07.817805 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:07 crc kubenswrapper[4767]: E0129 15:28:07.818414 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:07 crc kubenswrapper[4767]: E0129 15:28:07.818474 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert podName:dc435770-6862-4056-898e-ee9fbf4850d3 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:15.818456268 +0000 UTC m=+1651.628773307 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" (UID: "dc435770-6862-4056-898e-ee9fbf4850d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:08 crc kubenswrapper[4767]: I0129 15:28:08.122580 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:08 crc kubenswrapper[4767]: I0129 15:28:08.122676 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:08 crc kubenswrapper[4767]: E0129 15:28:08.122818 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 15:28:08 crc kubenswrapper[4767]: E0129 15:28:08.122886 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:16.122867165 +0000 UTC m=+1651.933184204 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "metrics-server-cert" not found Jan 29 15:28:08 crc kubenswrapper[4767]: E0129 15:28:08.123325 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 15:28:08 crc kubenswrapper[4767]: E0129 15:28:08.123754 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:16.123712049 +0000 UTC m=+1651.934029098 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "webhook-server-cert" not found Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.777962 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" event={"ID":"ca1f1109-544f-490f-9bf5-9d05241d30f2","Type":"ContainerStarted","Data":"724da768d26e1dc5fbbccde47fc24160d6f1ab0d1b460418305069e196e5efda"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.778577 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.780731 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" event={"ID":"a7f3205f-40e9-4eed-8305-4e0ca0ebe75c","Type":"ContainerStarted","Data":"313b173ddc4de8d8bca4fb4f26a0f2dbfa4e93651645ab00449819c36cf066ef"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.780812 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.782977 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" event={"ID":"a1b7dc5f-dccb-44e8-b390-daa6f9e773b3","Type":"ContainerStarted","Data":"5f58d4cb4b9597fe7c7b0508d7d5884f567a1930094363e7ffe335406282d04c"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.783705 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.785114 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" event={"ID":"7b712b0e-3913-4421-a6c8-90991284909a","Type":"ContainerStarted","Data":"cbe0195d0c4cd3fa41df57abf0bc227267ad920d92ad4d4616567719d42b5784"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.785608 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.787057 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" event={"ID":"49292974-81e4-4eba-8dc8-d6ac21111b9a","Type":"ContainerStarted","Data":"2db029ea254095d96a4f8aa560a46b6638a43a9e2d47ed72845fa8b88d3a8682"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.787518 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.789079 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" event={"ID":"68ca9f6a-ab9f-4188-a45c-23e6c1303a20","Type":"ContainerStarted","Data":"2bf6bff512f7fe22b152a75c1d255a655c5f7e87132aa8d5dbca41b965df580c"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.789543 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.791589 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" event={"ID":"833297bf-cdda-441d-8e03-de62a3b0c5b3","Type":"ContainerStarted","Data":"41bf7f2f253cbc3f77169ec5b2c1152ee05ea3470f9658b5d39abb725de30937"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.791747 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.793322 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" event={"ID":"adc494cd-a7d7-4104-9d5b-165115dfe504","Type":"ContainerStarted","Data":"195ff4d43e33ef0028df24fd26d667b7a7c3e53ec98c846cffe0e631d15d1ded"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.793452 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.796967 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" event={"ID":"6e272311-00bb-4cfb-92b5-2cf77eb483f8","Type":"ContainerStarted","Data":"de75967aadb75460ef2f7b79af5592d65a67bf221caafdc9459c9eae5d40c18a"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.797073 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.799772 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" event={"ID":"3350030b-3cea-4a51-9fdd-34408ad506eb","Type":"ContainerStarted","Data":"be953cc066080316d98ebf81eb1853b9f632359287827b94c411130c82817eb0"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.799916 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.801662 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" event={"ID":"a53807b1-819e-4e75-9f4b-ebda7fc7b415","Type":"ContainerStarted","Data":"27ed5a46503618ed66b37b4b420a10c04ee5e2afc90e5802dfb4d6545901b696"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.801825 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.803040 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" event={"ID":"157b230c-4bbc-4a62-841e-60acba588f1e","Type":"ContainerStarted","Data":"8c65ae156506b0869b1daf7f3df7c8cdde32cee0f3e640d0942349ac07937106"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.803386 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.804179 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" podStartSLOduration=3.4680507179999998 podStartE2EDuration="13.804161775s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.203268714 +0000 UTC m=+1637.013585753" lastFinishedPulling="2026-01-29 15:28:11.539379751 +0000 UTC m=+1647.349696810" observedRunningTime="2026-01-29 15:28:12.794970402 +0000 UTC m=+1648.605287441" watchObservedRunningTime="2026-01-29 15:28:12.804161775 +0000 UTC m=+1648.614478814" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.805039 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" event={"ID":"4be7d2d9-0fa7-4783-987a-7dd06823b118","Type":"ContainerStarted","Data":"de2b178e6d646d7797fc69a0414d7a8104c9e38adf50fef75ce4eb41442ccb6f"} Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.805153 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.805191 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.805245 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.822956 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" podStartSLOduration=3.709635025 podStartE2EDuration="13.822941202s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.429513872 +0000 UTC m=+1637.239830931" lastFinishedPulling="2026-01-29 15:28:11.542820059 +0000 UTC m=+1647.353137108" observedRunningTime="2026-01-29 15:28:12.820326338 +0000 UTC m=+1648.630643377" watchObservedRunningTime="2026-01-29 15:28:12.822941202 +0000 UTC m=+1648.633258241" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.846083 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" podStartSLOduration=3.545800234 podStartE2EDuration="13.846061844s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.200433003 +0000 UTC m=+1637.010750042" lastFinishedPulling="2026-01-29 15:28:11.500694593 +0000 UTC m=+1647.311011652" observedRunningTime="2026-01-29 15:28:12.841146054 +0000 UTC m=+1648.651463103" watchObservedRunningTime="2026-01-29 15:28:12.846061844 +0000 UTC m=+1648.656378883" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.854268 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" podStartSLOduration=3.492974382 podStartE2EDuration="13.854247939s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.198540309 +0000 UTC m=+1637.008857348" lastFinishedPulling="2026-01-29 15:28:11.559813866 +0000 UTC m=+1647.370130905" observedRunningTime="2026-01-29 15:28:12.85323306 +0000 UTC m=+1648.663550099" watchObservedRunningTime="2026-01-29 15:28:12.854247939 +0000 UTC m=+1648.664564968" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.872566 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" podStartSLOduration=3.448510339 podStartE2EDuration="13.872550363s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.077119342 +0000 UTC m=+1636.887436381" lastFinishedPulling="2026-01-29 15:28:11.501159366 +0000 UTC m=+1647.311476405" observedRunningTime="2026-01-29 15:28:12.871254536 +0000 UTC m=+1648.681571575" watchObservedRunningTime="2026-01-29 15:28:12.872550363 +0000 UTC m=+1648.682867402" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.899520 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" podStartSLOduration=3.505397888 podStartE2EDuration="13.899500125s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.19859135 +0000 UTC m=+1637.008908389" lastFinishedPulling="2026-01-29 15:28:11.592693577 +0000 UTC m=+1647.403010626" observedRunningTime="2026-01-29 15:28:12.899411112 +0000 UTC m=+1648.709728151" watchObservedRunningTime="2026-01-29 15:28:12.899500125 +0000 UTC m=+1648.709817164" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.923802 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" podStartSLOduration=3.761747288 podStartE2EDuration="13.92377714s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.379993994 +0000 UTC m=+1637.190311033" lastFinishedPulling="2026-01-29 15:28:11.542023846 +0000 UTC m=+1647.352340885" observedRunningTime="2026-01-29 15:28:12.920195757 +0000 UTC m=+1648.730512786" watchObservedRunningTime="2026-01-29 15:28:12.92377714 +0000 UTC m=+1648.734094179" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.950672 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" podStartSLOduration=3.510791112 podStartE2EDuration="13.950655179s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.055942856 +0000 UTC m=+1636.866259895" lastFinishedPulling="2026-01-29 15:28:11.495806903 +0000 UTC m=+1647.306123962" observedRunningTime="2026-01-29 15:28:12.946911142 +0000 UTC m=+1648.757228201" watchObservedRunningTime="2026-01-29 15:28:12.950655179 +0000 UTC m=+1648.760972218" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.963610 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" podStartSLOduration=3.464158777 podStartE2EDuration="13.9635945s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.024250328 +0000 UTC m=+1636.834567367" lastFinishedPulling="2026-01-29 15:28:11.523686031 +0000 UTC m=+1647.334003090" observedRunningTime="2026-01-29 15:28:12.95800592 +0000 UTC m=+1648.768322959" watchObservedRunningTime="2026-01-29 15:28:12.9635945 +0000 UTC m=+1648.773911539" Jan 29 15:28:12 crc kubenswrapper[4767]: I0129 15:28:12.980762 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" podStartSLOduration=3.140195751 podStartE2EDuration="13.980743601s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:00.678946271 +0000 UTC m=+1636.489263310" lastFinishedPulling="2026-01-29 15:28:11.519494121 +0000 UTC m=+1647.329811160" observedRunningTime="2026-01-29 15:28:12.977382645 +0000 UTC m=+1648.787699684" watchObservedRunningTime="2026-01-29 15:28:12.980743601 +0000 UTC m=+1648.791060640" Jan 29 15:28:13 crc kubenswrapper[4767]: I0129 15:28:13.005996 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" podStartSLOduration=3.6819013419999997 podStartE2EDuration="14.005976583s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.208628428 +0000 UTC m=+1637.018945467" lastFinishedPulling="2026-01-29 15:28:11.532703669 +0000 UTC m=+1647.343020708" observedRunningTime="2026-01-29 15:28:13.003219104 +0000 UTC m=+1648.813536143" watchObservedRunningTime="2026-01-29 15:28:13.005976583 +0000 UTC m=+1648.816293632" Jan 29 15:28:13 crc kubenswrapper[4767]: I0129 15:28:13.023645 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" podStartSLOduration=3.5764382120000002 podStartE2EDuration="14.023628919s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.072270783 +0000 UTC m=+1636.882587822" lastFinishedPulling="2026-01-29 15:28:11.51946148 +0000 UTC m=+1647.329778529" observedRunningTime="2026-01-29 15:28:13.022433035 +0000 UTC m=+1648.832750074" watchObservedRunningTime="2026-01-29 15:28:13.023628919 +0000 UTC m=+1648.833945948" Jan 29 15:28:13 crc kubenswrapper[4767]: I0129 15:28:13.047951 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" podStartSLOduration=3.606894884 podStartE2EDuration="14.047933975s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.096683892 +0000 UTC m=+1636.907000931" lastFinishedPulling="2026-01-29 15:28:11.537722983 +0000 UTC m=+1647.348040022" observedRunningTime="2026-01-29 15:28:13.04356211 +0000 UTC m=+1648.853879149" watchObservedRunningTime="2026-01-29 15:28:13.047933975 +0000 UTC m=+1648.858251014" Jan 29 15:28:15 crc kubenswrapper[4767]: I0129 15:28:15.438020 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:15 crc kubenswrapper[4767]: E0129 15:28:15.438285 4767 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:15 crc kubenswrapper[4767]: E0129 15:28:15.438458 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert podName:97c4098d-defd-4156-b48b-e26dd157f709 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:31.438439593 +0000 UTC m=+1667.248756632 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert") pod "infra-operator-controller-manager-79955696d6-br5bs" (UID: "97c4098d-defd-4156-b48b-e26dd157f709") : secret "infra-operator-webhook-server-cert" not found Jan 29 15:28:15 crc kubenswrapper[4767]: I0129 15:28:15.843820 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:15 crc kubenswrapper[4767]: E0129 15:28:15.844039 4767 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:15 crc kubenswrapper[4767]: E0129 15:28:15.844087 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert podName:dc435770-6862-4056-898e-ee9fbf4850d3 nodeName:}" failed. No retries permitted until 2026-01-29 15:28:31.844073907 +0000 UTC m=+1667.654390936 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert") pod "openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" (UID: "dc435770-6862-4056-898e-ee9fbf4850d3") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 29 15:28:16 crc kubenswrapper[4767]: I0129 15:28:16.148543 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:16 crc kubenswrapper[4767]: I0129 15:28:16.148741 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:16 crc kubenswrapper[4767]: E0129 15:28:16.149399 4767 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 29 15:28:16 crc kubenswrapper[4767]: E0129 15:28:16.149447 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:32.14942958 +0000 UTC m=+1667.959746639 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "metrics-server-cert" not found Jan 29 15:28:16 crc kubenswrapper[4767]: E0129 15:28:16.149802 4767 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 29 15:28:16 crc kubenswrapper[4767]: E0129 15:28:16.149835 4767 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs podName:b2a9e907-bdb8-408f-ba23-b13a5f792e5a nodeName:}" failed. No retries permitted until 2026-01-29 15:28:32.149825662 +0000 UTC m=+1667.960142711 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs") pod "openstack-operator-controller-manager-54c8dbf95-blw98" (UID: "b2a9e907-bdb8-408f-ba23-b13a5f792e5a") : secret "webhook-server-cert" not found Jan 29 15:28:19 crc kubenswrapper[4767]: I0129 15:28:19.671469 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d6fdb96dc-gkq4d" Jan 29 15:28:19 crc kubenswrapper[4767]: I0129 15:28:19.743098 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-858d89fd-v5bpj" Jan 29 15:28:19 crc kubenswrapper[4767]: I0129 15:28:19.770757 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-dd77988f8-bvxsd" Jan 29 15:28:19 crc kubenswrapper[4767]: I0129 15:28:19.797615 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-d8b84fbc-9zkzh" Jan 29 15:28:19 crc kubenswrapper[4767]: I0129 15:28:19.823806 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67bf948998-rlgl7" Jan 29 15:28:19 crc kubenswrapper[4767]: I0129 15:28:19.824670 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5fb775575f-6gtvq" Jan 29 15:28:19 crc kubenswrapper[4767]: I0129 15:28:19.911454 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c7cc6ff45-zwfl4" Jan 29 15:28:20 crc kubenswrapper[4767]: I0129 15:28:20.004263 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-866c9d5b98-hvtlx" Jan 29 15:28:20 crc kubenswrapper[4767]: I0129 15:28:20.004304 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-68cb478976-gbtpt" Jan 29 15:28:20 crc kubenswrapper[4767]: I0129 15:28:20.073306 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-8ccc8547b-gd2cz" Jan 29 15:28:20 crc kubenswrapper[4767]: I0129 15:28:20.255682 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-68f8cb846c-cs4g8" Jan 29 15:28:20 crc kubenswrapper[4767]: I0129 15:28:20.397962 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6f7455757b-pc4p9" Jan 29 15:28:20 crc kubenswrapper[4767]: I0129 15:28:20.557060 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56f8bfcd9f-dbphp" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.871143 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" event={"ID":"440d7f0e-a462-453a-b41f-ccdb1521bb20","Type":"ContainerStarted","Data":"3355c85d3a762c6aa07638948aa6aa2dd4953d5f24098099ff4775cab9da1a0f"} Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.871951 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.872991 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" event={"ID":"1a51e166-62bf-4181-b44f-62d52b543f2e","Type":"ContainerStarted","Data":"28d6e4dd7f5cc9e357d46f9f8daa6d17cb1d376e7344ca36ce444e332ce56c2b"} Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.873175 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.874980 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" event={"ID":"eba759c2-d8c7-4956-a47c-def368a54bf1","Type":"ContainerStarted","Data":"8ccca3ad48fb6dbc40161f6e943238641d2511c8d646c4b0d2f1584a0314f433"} Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.875347 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.876710 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" event={"ID":"0d956603-d824-432d-b1db-6574d2bde092","Type":"ContainerStarted","Data":"745dbb69c5ee452edbe9353103861b9010471233cdcdef20281d2e62ea6eb17b"} Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.877111 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.878549 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" event={"ID":"f104967e-4b06-42b6-b9ea-c4edb1923e87","Type":"ContainerStarted","Data":"b3b5717a65375c029182cd975b5690d33e58ea8eafc69857de0e39f649e72dc4"} Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.880343 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" event={"ID":"b186845a-5b3b-42ce-aad0-0269580f5b07","Type":"ContainerStarted","Data":"fc2a3302956e98d0073c47ef1ea3e08a84548431546c0f538ceb769c8e3c5a36"} Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.880510 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.881907 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" event={"ID":"e63115ec-b52a-4e9f-89b5-76004e0f6316","Type":"ContainerStarted","Data":"8fad0a27f01fbcbf11faba62c2a3453b214a08386804bba8c66bb18487312209"} Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.882046 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.921931 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" podStartSLOduration=3.119109887 podStartE2EDuration="22.921912596s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.415535392 +0000 UTC m=+1637.225852431" lastFinishedPulling="2026-01-29 15:28:21.218338101 +0000 UTC m=+1657.028655140" observedRunningTime="2026-01-29 15:28:21.915375879 +0000 UTC m=+1657.725692918" watchObservedRunningTime="2026-01-29 15:28:21.921912596 +0000 UTC m=+1657.732229635" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.939310 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" podStartSLOduration=3.157030554 podStartE2EDuration="22.939292774s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.39588883 +0000 UTC m=+1637.206205879" lastFinishedPulling="2026-01-29 15:28:21.17815106 +0000 UTC m=+1656.988468099" observedRunningTime="2026-01-29 15:28:21.935884897 +0000 UTC m=+1657.746201936" watchObservedRunningTime="2026-01-29 15:28:21.939292774 +0000 UTC m=+1657.749609803" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.953426 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" podStartSLOduration=2.946904117 podStartE2EDuration="22.953406618s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.21186205 +0000 UTC m=+1637.022179079" lastFinishedPulling="2026-01-29 15:28:21.218364541 +0000 UTC m=+1657.028681580" observedRunningTime="2026-01-29 15:28:21.951237716 +0000 UTC m=+1657.761554755" watchObservedRunningTime="2026-01-29 15:28:21.953406618 +0000 UTC m=+1657.763723657" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.975056 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" podStartSLOduration=3.173472385 podStartE2EDuration="22.975035078s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.385491532 +0000 UTC m=+1637.195808571" lastFinishedPulling="2026-01-29 15:28:21.187054225 +0000 UTC m=+1656.997371264" observedRunningTime="2026-01-29 15:28:21.971683572 +0000 UTC m=+1657.782000611" watchObservedRunningTime="2026-01-29 15:28:21.975035078 +0000 UTC m=+1657.785352127" Jan 29 15:28:21 crc kubenswrapper[4767]: I0129 15:28:21.990880 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-xmjrr" podStartSLOduration=3.3637232360000002 podStartE2EDuration="21.990860721s" podCreationTimestamp="2026-01-29 15:28:00 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.396064215 +0000 UTC m=+1637.206381254" lastFinishedPulling="2026-01-29 15:28:20.02320169 +0000 UTC m=+1655.833518739" observedRunningTime="2026-01-29 15:28:21.98839861 +0000 UTC m=+1657.798715659" watchObservedRunningTime="2026-01-29 15:28:21.990860721 +0000 UTC m=+1657.801177770" Jan 29 15:28:22 crc kubenswrapper[4767]: I0129 15:28:22.014299 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" podStartSLOduration=3.215027024 podStartE2EDuration="23.014283621s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.38856389 +0000 UTC m=+1637.198880929" lastFinishedPulling="2026-01-29 15:28:21.187820487 +0000 UTC m=+1656.998137526" observedRunningTime="2026-01-29 15:28:22.010604386 +0000 UTC m=+1657.820921425" watchObservedRunningTime="2026-01-29 15:28:22.014283621 +0000 UTC m=+1657.824600660" Jan 29 15:28:29 crc kubenswrapper[4767]: I0129 15:28:29.774740 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-f8c4db9df-r4nvx" Jan 29 15:28:29 crc kubenswrapper[4767]: I0129 15:28:29.796382 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" podStartSLOduration=12.172387244 podStartE2EDuration="30.796363639s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:01.393647205 +0000 UTC m=+1637.203964244" lastFinishedPulling="2026-01-29 15:28:20.0176236 +0000 UTC m=+1655.827940639" observedRunningTime="2026-01-29 15:28:22.031677459 +0000 UTC m=+1657.841994498" watchObservedRunningTime="2026-01-29 15:28:29.796363639 +0000 UTC m=+1665.606680688" Jan 29 15:28:29 crc kubenswrapper[4767]: I0129 15:28:29.961544 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-76c896469f-xxrwp" Jan 29 15:28:30 crc kubenswrapper[4767]: I0129 15:28:30.273610 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-788c46999f-gbkxz" Jan 29 15:28:30 crc kubenswrapper[4767]: I0129 15:28:30.346584 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5b964cf4cd-fsgxh" Jan 29 15:28:30 crc kubenswrapper[4767]: I0129 15:28:30.510019 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6cf8c44c7-nzz8p" Jan 29 15:28:30 crc kubenswrapper[4767]: I0129 15:28:30.599069 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-59f4c7d7c4-g6hsv" Jan 29 15:28:31 crc kubenswrapper[4767]: I0129 15:28:31.509820 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:31 crc kubenswrapper[4767]: I0129 15:28:31.517096 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/97c4098d-defd-4156-b48b-e26dd157f709-cert\") pod \"infra-operator-controller-manager-79955696d6-br5bs\" (UID: \"97c4098d-defd-4156-b48b-e26dd157f709\") " pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:31 crc kubenswrapper[4767]: I0129 15:28:31.656520 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:31 crc kubenswrapper[4767]: I0129 15:28:31.915635 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:31 crc kubenswrapper[4767]: I0129 15:28:31.922797 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dc435770-6862-4056-898e-ee9fbf4850d3-cert\") pod \"openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4\" (UID: \"dc435770-6862-4056-898e-ee9fbf4850d3\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.074968 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79955696d6-br5bs"] Jan 29 15:28:32 crc kubenswrapper[4767]: W0129 15:28:32.079850 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97c4098d_defd_4156_b48b_e26dd157f709.slice/crio-40c2e64fd687be26eb30f300f59637732db82426e3168ff7fd896a02d870b07f WatchSource:0}: Error finding container 40c2e64fd687be26eb30f300f59637732db82426e3168ff7fd896a02d870b07f: Status 404 returned error can't find the container with id 40c2e64fd687be26eb30f300f59637732db82426e3168ff7fd896a02d870b07f Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.090301 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.222428 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.222510 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.226304 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-webhook-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.226454 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2a9e907-bdb8-408f-ba23-b13a5f792e5a-metrics-certs\") pod \"openstack-operator-controller-manager-54c8dbf95-blw98\" (UID: \"b2a9e907-bdb8-408f-ba23-b13a5f792e5a\") " pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.313385 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4"] Jan 29 15:28:32 crc kubenswrapper[4767]: W0129 15:28:32.319872 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc435770_6862_4056_898e_ee9fbf4850d3.slice/crio-3ed4966abd04e0fdc505ce544d24337990398147d823e520221040cd74263470 WatchSource:0}: Error finding container 3ed4966abd04e0fdc505ce544d24337990398147d823e520221040cd74263470: Status 404 returned error can't find the container with id 3ed4966abd04e0fdc505ce544d24337990398147d823e520221040cd74263470 Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.504778 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.774121 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98"] Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.995608 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" event={"ID":"b2a9e907-bdb8-408f-ba23-b13a5f792e5a","Type":"ContainerStarted","Data":"3deb3f73c881b841ba5fee25559af1ece708f92992887ff613e05fa3b972f2a2"} Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.996916 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" event={"ID":"dc435770-6862-4056-898e-ee9fbf4850d3","Type":"ContainerStarted","Data":"3ed4966abd04e0fdc505ce544d24337990398147d823e520221040cd74263470"} Jan 29 15:28:32 crc kubenswrapper[4767]: I0129 15:28:32.998052 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" event={"ID":"97c4098d-defd-4156-b48b-e26dd157f709","Type":"ContainerStarted","Data":"40c2e64fd687be26eb30f300f59637732db82426e3168ff7fd896a02d870b07f"} Jan 29 15:28:34 crc kubenswrapper[4767]: I0129 15:28:34.009791 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" event={"ID":"b2a9e907-bdb8-408f-ba23-b13a5f792e5a","Type":"ContainerStarted","Data":"e391e28bff16ceda0957f2575b473cb402b51e21e6fbd6d1c9f5d6714e73c088"} Jan 29 15:28:34 crc kubenswrapper[4767]: I0129 15:28:34.011315 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:34 crc kubenswrapper[4767]: I0129 15:28:34.064002 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" podStartSLOduration=34.063982044 podStartE2EDuration="34.063982044s" podCreationTimestamp="2026-01-29 15:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-29 15:28:34.057357775 +0000 UTC m=+1669.867674864" watchObservedRunningTime="2026-01-29 15:28:34.063982044 +0000 UTC m=+1669.874299093" Jan 29 15:28:38 crc kubenswrapper[4767]: I0129 15:28:38.048882 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" event={"ID":"dc435770-6862-4056-898e-ee9fbf4850d3","Type":"ContainerStarted","Data":"172d06db893fd09a6642fc4a6bd5ad4a7c47e63c37432cc00e3befdaadd921b3"} Jan 29 15:28:38 crc kubenswrapper[4767]: I0129 15:28:38.049551 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:38 crc kubenswrapper[4767]: I0129 15:28:38.051825 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" event={"ID":"97c4098d-defd-4156-b48b-e26dd157f709","Type":"ContainerStarted","Data":"486273e923afea71ad4e4c900e4ade6f83abac5b8540f78401106811198e1041"} Jan 29 15:28:38 crc kubenswrapper[4767]: I0129 15:28:38.052005 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:38 crc kubenswrapper[4767]: I0129 15:28:38.083518 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" podStartSLOduration=34.21867427 podStartE2EDuration="39.083495266s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:32.322053848 +0000 UTC m=+1668.132370887" lastFinishedPulling="2026-01-29 15:28:37.186874844 +0000 UTC m=+1672.997191883" observedRunningTime="2026-01-29 15:28:38.076612109 +0000 UTC m=+1673.886929148" watchObservedRunningTime="2026-01-29 15:28:38.083495266 +0000 UTC m=+1673.893812315" Jan 29 15:28:38 crc kubenswrapper[4767]: I0129 15:28:38.111245 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" podStartSLOduration=34.006044362 podStartE2EDuration="39.1112175s" podCreationTimestamp="2026-01-29 15:27:59 +0000 UTC" firstStartedPulling="2026-01-29 15:28:32.082690944 +0000 UTC m=+1667.893007973" lastFinishedPulling="2026-01-29 15:28:37.187864072 +0000 UTC m=+1672.998181111" observedRunningTime="2026-01-29 15:28:38.096782737 +0000 UTC m=+1673.907099796" watchObservedRunningTime="2026-01-29 15:28:38.1112175 +0000 UTC m=+1673.921534589" Jan 29 15:28:42 crc kubenswrapper[4767]: I0129 15:28:42.102307 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4" Jan 29 15:28:42 crc kubenswrapper[4767]: I0129 15:28:42.513636 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-54c8dbf95-blw98" Jan 29 15:28:42 crc kubenswrapper[4767]: I0129 15:28:42.806132 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:28:42 crc kubenswrapper[4767]: I0129 15:28:42.806223 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:28:42 crc kubenswrapper[4767]: I0129 15:28:42.806278 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:28:42 crc kubenswrapper[4767]: I0129 15:28:42.807055 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:28:42 crc kubenswrapper[4767]: I0129 15:28:42.807122 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" gracePeriod=600 Jan 29 15:28:42 crc kubenswrapper[4767]: E0129 15:28:42.934217 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:28:43 crc kubenswrapper[4767]: I0129 15:28:43.091853 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" exitCode=0 Jan 29 15:28:43 crc kubenswrapper[4767]: I0129 15:28:43.091967 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce"} Jan 29 15:28:43 crc kubenswrapper[4767]: I0129 15:28:43.092311 4767 scope.go:117] "RemoveContainer" containerID="20bafaf87627c386be9cca699d2328988c5557f20c2026d00c954ee5e4a7303e" Jan 29 15:28:43 crc kubenswrapper[4767]: I0129 15:28:43.092835 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:28:43 crc kubenswrapper[4767]: E0129 15:28:43.093108 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:28:51 crc kubenswrapper[4767]: I0129 15:28:51.662940 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79955696d6-br5bs" Jan 29 15:28:55 crc kubenswrapper[4767]: I0129 15:28:55.026357 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:28:55 crc kubenswrapper[4767]: E0129 15:28:55.026974 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:29:00 crc kubenswrapper[4767]: I0129 15:29:00.941760 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d4n6x"] Jan 29 15:29:00 crc kubenswrapper[4767]: I0129 15:29:00.943842 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:00 crc kubenswrapper[4767]: I0129 15:29:00.954487 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4n6x"] Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.035405 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lfc9\" (UniqueName: \"kubernetes.io/projected/37aafec5-ae7f-47d5-a90f-e5bff4004051-kube-api-access-7lfc9\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.035450 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-utilities\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.035495 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-catalog-content\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.136953 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lfc9\" (UniqueName: \"kubernetes.io/projected/37aafec5-ae7f-47d5-a90f-e5bff4004051-kube-api-access-7lfc9\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.137008 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-utilities\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.137041 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-catalog-content\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.139118 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-utilities\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.139340 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-catalog-content\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.159630 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lfc9\" (UniqueName: \"kubernetes.io/projected/37aafec5-ae7f-47d5-a90f-e5bff4004051-kube-api-access-7lfc9\") pod \"redhat-operators-d4n6x\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.264333 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:01 crc kubenswrapper[4767]: I0129 15:29:01.502980 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d4n6x"] Jan 29 15:29:02 crc kubenswrapper[4767]: I0129 15:29:02.227678 4767 generic.go:334] "Generic (PLEG): container finished" podID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerID="5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec" exitCode=0 Jan 29 15:29:02 crc kubenswrapper[4767]: I0129 15:29:02.227866 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4n6x" event={"ID":"37aafec5-ae7f-47d5-a90f-e5bff4004051","Type":"ContainerDied","Data":"5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec"} Jan 29 15:29:02 crc kubenswrapper[4767]: I0129 15:29:02.229037 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4n6x" event={"ID":"37aafec5-ae7f-47d5-a90f-e5bff4004051","Type":"ContainerStarted","Data":"da04598bef560f9bd9154cb30bccc3e5522cf5e4af18b7a5273b9691e515d506"} Jan 29 15:29:02 crc kubenswrapper[4767]: E0129 15:29:02.353078 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:29:02 crc kubenswrapper[4767]: E0129 15:29:02.353306 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7lfc9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-d4n6x_openshift-marketplace(37aafec5-ae7f-47d5-a90f-e5bff4004051): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:29:02 crc kubenswrapper[4767]: E0129 15:29:02.354531 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-d4n6x" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" Jan 29 15:29:03 crc kubenswrapper[4767]: E0129 15:29:03.238195 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-d4n6x" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" Jan 29 15:29:06 crc kubenswrapper[4767]: I0129 15:29:06.019050 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:29:06 crc kubenswrapper[4767]: E0129 15:29:06.019554 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:29:18 crc kubenswrapper[4767]: I0129 15:29:18.018846 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:29:18 crc kubenswrapper[4767]: E0129 15:29:18.019540 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:29:19 crc kubenswrapper[4767]: I0129 15:29:19.355323 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4n6x" event={"ID":"37aafec5-ae7f-47d5-a90f-e5bff4004051","Type":"ContainerStarted","Data":"cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e"} Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.205438 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n9ncs"] Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.207508 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.219804 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9ncs"] Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.336342 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-utilities\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.336410 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-catalog-content\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.336441 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9dzd\" (UniqueName: \"kubernetes.io/projected/0a05bab3-5695-4b94-adc8-c116e482cdeb-kube-api-access-j9dzd\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.363772 4767 generic.go:334] "Generic (PLEG): container finished" podID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerID="cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e" exitCode=0 Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.363818 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4n6x" event={"ID":"37aafec5-ae7f-47d5-a90f-e5bff4004051","Type":"ContainerDied","Data":"cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e"} Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.437750 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-catalog-content\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.438355 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9dzd\" (UniqueName: \"kubernetes.io/projected/0a05bab3-5695-4b94-adc8-c116e482cdeb-kube-api-access-j9dzd\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.438278 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-catalog-content\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.439501 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-utilities\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.439541 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-utilities\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.457827 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9dzd\" (UniqueName: \"kubernetes.io/projected/0a05bab3-5695-4b94-adc8-c116e482cdeb-kube-api-access-j9dzd\") pod \"redhat-marketplace-n9ncs\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.529624 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:20 crc kubenswrapper[4767]: I0129 15:29:20.970664 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9ncs"] Jan 29 15:29:20 crc kubenswrapper[4767]: W0129 15:29:20.985945 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a05bab3_5695_4b94_adc8_c116e482cdeb.slice/crio-ddc438bc6b594fd7e537a0a474ea4d4b6f433f96e2e55a3d77fa056b1db60e13 WatchSource:0}: Error finding container ddc438bc6b594fd7e537a0a474ea4d4b6f433f96e2e55a3d77fa056b1db60e13: Status 404 returned error can't find the container with id ddc438bc6b594fd7e537a0a474ea4d4b6f433f96e2e55a3d77fa056b1db60e13 Jan 29 15:29:21 crc kubenswrapper[4767]: I0129 15:29:21.374747 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4n6x" event={"ID":"37aafec5-ae7f-47d5-a90f-e5bff4004051","Type":"ContainerStarted","Data":"0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954"} Jan 29 15:29:21 crc kubenswrapper[4767]: I0129 15:29:21.376271 4767 generic.go:334] "Generic (PLEG): container finished" podID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerID="cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461" exitCode=0 Jan 29 15:29:21 crc kubenswrapper[4767]: I0129 15:29:21.376297 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9ncs" event={"ID":"0a05bab3-5695-4b94-adc8-c116e482cdeb","Type":"ContainerDied","Data":"cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461"} Jan 29 15:29:21 crc kubenswrapper[4767]: I0129 15:29:21.376311 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9ncs" event={"ID":"0a05bab3-5695-4b94-adc8-c116e482cdeb","Type":"ContainerStarted","Data":"ddc438bc6b594fd7e537a0a474ea4d4b6f433f96e2e55a3d77fa056b1db60e13"} Jan 29 15:29:21 crc kubenswrapper[4767]: I0129 15:29:21.393791 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d4n6x" podStartSLOduration=2.858324773 podStartE2EDuration="21.393766464s" podCreationTimestamp="2026-01-29 15:29:00 +0000 UTC" firstStartedPulling="2026-01-29 15:29:02.229322792 +0000 UTC m=+1698.039639831" lastFinishedPulling="2026-01-29 15:29:20.764764493 +0000 UTC m=+1716.575081522" observedRunningTime="2026-01-29 15:29:21.391244162 +0000 UTC m=+1717.201561211" watchObservedRunningTime="2026-01-29 15:29:21.393766464 +0000 UTC m=+1717.204083513" Jan 29 15:29:22 crc kubenswrapper[4767]: I0129 15:29:22.384294 4767 generic.go:334] "Generic (PLEG): container finished" podID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerID="253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290" exitCode=0 Jan 29 15:29:22 crc kubenswrapper[4767]: I0129 15:29:22.384347 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9ncs" event={"ID":"0a05bab3-5695-4b94-adc8-c116e482cdeb","Type":"ContainerDied","Data":"253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290"} Jan 29 15:29:23 crc kubenswrapper[4767]: I0129 15:29:23.392158 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9ncs" event={"ID":"0a05bab3-5695-4b94-adc8-c116e482cdeb","Type":"ContainerStarted","Data":"7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903"} Jan 29 15:29:23 crc kubenswrapper[4767]: I0129 15:29:23.415316 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n9ncs" podStartSLOduration=2.049337854 podStartE2EDuration="3.415298687s" podCreationTimestamp="2026-01-29 15:29:20 +0000 UTC" firstStartedPulling="2026-01-29 15:29:21.379010851 +0000 UTC m=+1717.189327890" lastFinishedPulling="2026-01-29 15:29:22.744971684 +0000 UTC m=+1718.555288723" observedRunningTime="2026-01-29 15:29:23.41190978 +0000 UTC m=+1719.222226819" watchObservedRunningTime="2026-01-29 15:29:23.415298687 +0000 UTC m=+1719.225615746" Jan 29 15:29:29 crc kubenswrapper[4767]: I0129 15:29:29.018672 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:29:29 crc kubenswrapper[4767]: E0129 15:29:29.019491 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:29:30 crc kubenswrapper[4767]: I0129 15:29:30.530471 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:30 crc kubenswrapper[4767]: I0129 15:29:30.530834 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:30 crc kubenswrapper[4767]: I0129 15:29:30.590955 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:31 crc kubenswrapper[4767]: I0129 15:29:31.265165 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:31 crc kubenswrapper[4767]: I0129 15:29:31.265497 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:31 crc kubenswrapper[4767]: I0129 15:29:31.307536 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:31 crc kubenswrapper[4767]: I0129 15:29:31.494524 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:31 crc kubenswrapper[4767]: I0129 15:29:31.503245 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:32 crc kubenswrapper[4767]: I0129 15:29:32.031772 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4n6x"] Jan 29 15:29:33 crc kubenswrapper[4767]: I0129 15:29:33.464749 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d4n6x" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="registry-server" containerID="cri-o://0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954" gracePeriod=2 Jan 29 15:29:33 crc kubenswrapper[4767]: I0129 15:29:33.832164 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9ncs"] Jan 29 15:29:33 crc kubenswrapper[4767]: I0129 15:29:33.832653 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n9ncs" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="registry-server" containerID="cri-o://7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903" gracePeriod=2 Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.219503 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.341655 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9dzd\" (UniqueName: \"kubernetes.io/projected/0a05bab3-5695-4b94-adc8-c116e482cdeb-kube-api-access-j9dzd\") pod \"0a05bab3-5695-4b94-adc8-c116e482cdeb\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.341722 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-catalog-content\") pod \"0a05bab3-5695-4b94-adc8-c116e482cdeb\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.341755 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-utilities\") pod \"0a05bab3-5695-4b94-adc8-c116e482cdeb\" (UID: \"0a05bab3-5695-4b94-adc8-c116e482cdeb\") " Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.342986 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-utilities" (OuterVolumeSpecName: "utilities") pod "0a05bab3-5695-4b94-adc8-c116e482cdeb" (UID: "0a05bab3-5695-4b94-adc8-c116e482cdeb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.349037 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a05bab3-5695-4b94-adc8-c116e482cdeb-kube-api-access-j9dzd" (OuterVolumeSpecName: "kube-api-access-j9dzd") pod "0a05bab3-5695-4b94-adc8-c116e482cdeb" (UID: "0a05bab3-5695-4b94-adc8-c116e482cdeb"). InnerVolumeSpecName "kube-api-access-j9dzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.362384 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a05bab3-5695-4b94-adc8-c116e482cdeb" (UID: "0a05bab3-5695-4b94-adc8-c116e482cdeb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.394449 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.443521 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-catalog-content\") pod \"37aafec5-ae7f-47d5-a90f-e5bff4004051\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.443719 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lfc9\" (UniqueName: \"kubernetes.io/projected/37aafec5-ae7f-47d5-a90f-e5bff4004051-kube-api-access-7lfc9\") pod \"37aafec5-ae7f-47d5-a90f-e5bff4004051\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.443805 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-utilities\") pod \"37aafec5-ae7f-47d5-a90f-e5bff4004051\" (UID: \"37aafec5-ae7f-47d5-a90f-e5bff4004051\") " Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.444312 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9dzd\" (UniqueName: \"kubernetes.io/projected/0a05bab3-5695-4b94-adc8-c116e482cdeb-kube-api-access-j9dzd\") on node \"crc\" DevicePath \"\"" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.444348 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.444369 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a05bab3-5695-4b94-adc8-c116e482cdeb-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.445060 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-utilities" (OuterVolumeSpecName: "utilities") pod "37aafec5-ae7f-47d5-a90f-e5bff4004051" (UID: "37aafec5-ae7f-47d5-a90f-e5bff4004051"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.447232 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37aafec5-ae7f-47d5-a90f-e5bff4004051-kube-api-access-7lfc9" (OuterVolumeSpecName: "kube-api-access-7lfc9") pod "37aafec5-ae7f-47d5-a90f-e5bff4004051" (UID: "37aafec5-ae7f-47d5-a90f-e5bff4004051"). InnerVolumeSpecName "kube-api-access-7lfc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.472650 4767 generic.go:334] "Generic (PLEG): container finished" podID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerID="0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954" exitCode=0 Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.472708 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4n6x" event={"ID":"37aafec5-ae7f-47d5-a90f-e5bff4004051","Type":"ContainerDied","Data":"0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954"} Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.472754 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d4n6x" event={"ID":"37aafec5-ae7f-47d5-a90f-e5bff4004051","Type":"ContainerDied","Data":"da04598bef560f9bd9154cb30bccc3e5522cf5e4af18b7a5273b9691e515d506"} Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.472771 4767 scope.go:117] "RemoveContainer" containerID="0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.472943 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d4n6x" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.477227 4767 generic.go:334] "Generic (PLEG): container finished" podID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerID="7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903" exitCode=0 Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.477278 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9ncs" event={"ID":"0a05bab3-5695-4b94-adc8-c116e482cdeb","Type":"ContainerDied","Data":"7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903"} Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.477312 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n9ncs" event={"ID":"0a05bab3-5695-4b94-adc8-c116e482cdeb","Type":"ContainerDied","Data":"ddc438bc6b594fd7e537a0a474ea4d4b6f433f96e2e55a3d77fa056b1db60e13"} Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.477332 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n9ncs" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.495156 4767 scope.go:117] "RemoveContainer" containerID="cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.508676 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9ncs"] Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.514640 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n9ncs"] Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.523968 4767 scope.go:117] "RemoveContainer" containerID="5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.542261 4767 scope.go:117] "RemoveContainer" containerID="0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954" Jan 29 15:29:34 crc kubenswrapper[4767]: E0129 15:29:34.542797 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954\": container with ID starting with 0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954 not found: ID does not exist" containerID="0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.542842 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954"} err="failed to get container status \"0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954\": rpc error: code = NotFound desc = could not find container \"0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954\": container with ID starting with 0894d63e344aaf1aa2f426ad0ae3beecc66318b30b90617460d7b42cd2e41954 not found: ID does not exist" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.542874 4767 scope.go:117] "RemoveContainer" containerID="cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e" Jan 29 15:29:34 crc kubenswrapper[4767]: E0129 15:29:34.543287 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e\": container with ID starting with cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e not found: ID does not exist" containerID="cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.543316 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e"} err="failed to get container status \"cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e\": rpc error: code = NotFound desc = could not find container \"cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e\": container with ID starting with cd722a4326173aa2eed92e57391c6d0e2a21f2bc4abbeb3bd5c7eceb888ad90e not found: ID does not exist" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.543341 4767 scope.go:117] "RemoveContainer" containerID="5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec" Jan 29 15:29:34 crc kubenswrapper[4767]: E0129 15:29:34.543578 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec\": container with ID starting with 5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec not found: ID does not exist" containerID="5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.543601 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec"} err="failed to get container status \"5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec\": rpc error: code = NotFound desc = could not find container \"5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec\": container with ID starting with 5a7de8bf67aa2917f7073a01af69c7bdceaee68be0d08cf3b8a173bedf66a6ec not found: ID does not exist" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.543615 4767 scope.go:117] "RemoveContainer" containerID="7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.545073 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lfc9\" (UniqueName: \"kubernetes.io/projected/37aafec5-ae7f-47d5-a90f-e5bff4004051-kube-api-access-7lfc9\") on node \"crc\" DevicePath \"\"" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.545093 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.558580 4767 scope.go:117] "RemoveContainer" containerID="253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.570194 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37aafec5-ae7f-47d5-a90f-e5bff4004051" (UID: "37aafec5-ae7f-47d5-a90f-e5bff4004051"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.574711 4767 scope.go:117] "RemoveContainer" containerID="cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.618386 4767 scope.go:117] "RemoveContainer" containerID="7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903" Jan 29 15:29:34 crc kubenswrapper[4767]: E0129 15:29:34.618961 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903\": container with ID starting with 7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903 not found: ID does not exist" containerID="7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.619024 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903"} err="failed to get container status \"7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903\": rpc error: code = NotFound desc = could not find container \"7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903\": container with ID starting with 7a5e403c2606f17a82ea67c5bb8b4fc6dcdb45cff443bda780317eecd66fa903 not found: ID does not exist" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.619056 4767 scope.go:117] "RemoveContainer" containerID="253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290" Jan 29 15:29:34 crc kubenswrapper[4767]: E0129 15:29:34.619591 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290\": container with ID starting with 253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290 not found: ID does not exist" containerID="253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.619626 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290"} err="failed to get container status \"253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290\": rpc error: code = NotFound desc = could not find container \"253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290\": container with ID starting with 253d08e8c6b39bc85bcfeded01d53f7e31a0f8635d2f4a35c2a9d417dc9ba290 not found: ID does not exist" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.619672 4767 scope.go:117] "RemoveContainer" containerID="cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461" Jan 29 15:29:34 crc kubenswrapper[4767]: E0129 15:29:34.619968 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461\": container with ID starting with cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461 not found: ID does not exist" containerID="cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.619991 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461"} err="failed to get container status \"cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461\": rpc error: code = NotFound desc = could not find container \"cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461\": container with ID starting with cf25b62e4555acda8fe5eb8588f68835133a94d29d75e87c028199effa76f461 not found: ID does not exist" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.646011 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37aafec5-ae7f-47d5-a90f-e5bff4004051-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.810753 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d4n6x"] Jan 29 15:29:34 crc kubenswrapper[4767]: I0129 15:29:34.817025 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d4n6x"] Jan 29 15:29:35 crc kubenswrapper[4767]: I0129 15:29:35.037872 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" path="/var/lib/kubelet/pods/0a05bab3-5695-4b94-adc8-c116e482cdeb/volumes" Jan 29 15:29:35 crc kubenswrapper[4767]: I0129 15:29:35.038491 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" path="/var/lib/kubelet/pods/37aafec5-ae7f-47d5-a90f-e5bff4004051/volumes" Jan 29 15:29:41 crc kubenswrapper[4767]: I0129 15:29:41.018815 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:29:41 crc kubenswrapper[4767]: E0129 15:29:41.019549 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:29:56 crc kubenswrapper[4767]: I0129 15:29:56.019437 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:29:56 crc kubenswrapper[4767]: E0129 15:29:56.020157 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.151754 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h"] Jan 29 15:30:00 crc kubenswrapper[4767]: E0129 15:30:00.152498 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="registry-server" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152517 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="registry-server" Jan 29 15:30:00 crc kubenswrapper[4767]: E0129 15:30:00.152546 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="extract-content" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152555 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="extract-content" Jan 29 15:30:00 crc kubenswrapper[4767]: E0129 15:30:00.152591 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="registry-server" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152603 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="registry-server" Jan 29 15:30:00 crc kubenswrapper[4767]: E0129 15:30:00.152615 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="extract-content" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152624 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="extract-content" Jan 29 15:30:00 crc kubenswrapper[4767]: E0129 15:30:00.152634 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="extract-utilities" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152643 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="extract-utilities" Jan 29 15:30:00 crc kubenswrapper[4767]: E0129 15:30:00.152667 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="extract-utilities" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152675 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="extract-utilities" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152826 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a05bab3-5695-4b94-adc8-c116e482cdeb" containerName="registry-server" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.152854 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="37aafec5-ae7f-47d5-a90f-e5bff4004051" containerName="registry-server" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.153549 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.155878 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.157168 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.163659 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h"] Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.221838 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0f41608-5a18-4065-90cb-2c66870cb435-config-volume\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.222303 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g7hs\" (UniqueName: \"kubernetes.io/projected/f0f41608-5a18-4065-90cb-2c66870cb435-kube-api-access-8g7hs\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.222333 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f0f41608-5a18-4065-90cb-2c66870cb435-secret-volume\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.323968 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g7hs\" (UniqueName: \"kubernetes.io/projected/f0f41608-5a18-4065-90cb-2c66870cb435-kube-api-access-8g7hs\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.324032 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f0f41608-5a18-4065-90cb-2c66870cb435-secret-volume\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.324064 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0f41608-5a18-4065-90cb-2c66870cb435-config-volume\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.325064 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0f41608-5a18-4065-90cb-2c66870cb435-config-volume\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.331152 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f0f41608-5a18-4065-90cb-2c66870cb435-secret-volume\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.342052 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g7hs\" (UniqueName: \"kubernetes.io/projected/f0f41608-5a18-4065-90cb-2c66870cb435-kube-api-access-8g7hs\") pod \"collect-profiles-29495010-hg69h\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.473815 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:00 crc kubenswrapper[4767]: I0129 15:30:00.877140 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h"] Jan 29 15:30:01 crc kubenswrapper[4767]: I0129 15:30:01.684125 4767 generic.go:334] "Generic (PLEG): container finished" podID="f0f41608-5a18-4065-90cb-2c66870cb435" containerID="1fff907142bbc1b700498a8835f6c96f0705cfd4655d46347bbcca0d10e239ac" exitCode=0 Jan 29 15:30:01 crc kubenswrapper[4767]: I0129 15:30:01.684174 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" event={"ID":"f0f41608-5a18-4065-90cb-2c66870cb435","Type":"ContainerDied","Data":"1fff907142bbc1b700498a8835f6c96f0705cfd4655d46347bbcca0d10e239ac"} Jan 29 15:30:01 crc kubenswrapper[4767]: I0129 15:30:01.684226 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" event={"ID":"f0f41608-5a18-4065-90cb-2c66870cb435","Type":"ContainerStarted","Data":"57187ac322569d1d4b36fe886497f108760f9afc90f2413a67a10822a2ed5441"} Jan 29 15:30:02 crc kubenswrapper[4767]: I0129 15:30:02.987539 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.060701 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g7hs\" (UniqueName: \"kubernetes.io/projected/f0f41608-5a18-4065-90cb-2c66870cb435-kube-api-access-8g7hs\") pod \"f0f41608-5a18-4065-90cb-2c66870cb435\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.060758 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f0f41608-5a18-4065-90cb-2c66870cb435-secret-volume\") pod \"f0f41608-5a18-4065-90cb-2c66870cb435\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.060831 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0f41608-5a18-4065-90cb-2c66870cb435-config-volume\") pod \"f0f41608-5a18-4065-90cb-2c66870cb435\" (UID: \"f0f41608-5a18-4065-90cb-2c66870cb435\") " Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.061725 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0f41608-5a18-4065-90cb-2c66870cb435-config-volume" (OuterVolumeSpecName: "config-volume") pod "f0f41608-5a18-4065-90cb-2c66870cb435" (UID: "f0f41608-5a18-4065-90cb-2c66870cb435"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.066306 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0f41608-5a18-4065-90cb-2c66870cb435-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f0f41608-5a18-4065-90cb-2c66870cb435" (UID: "f0f41608-5a18-4065-90cb-2c66870cb435"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.066328 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0f41608-5a18-4065-90cb-2c66870cb435-kube-api-access-8g7hs" (OuterVolumeSpecName: "kube-api-access-8g7hs") pod "f0f41608-5a18-4065-90cb-2c66870cb435" (UID: "f0f41608-5a18-4065-90cb-2c66870cb435"). InnerVolumeSpecName "kube-api-access-8g7hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.162544 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f0f41608-5a18-4065-90cb-2c66870cb435-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.162861 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g7hs\" (UniqueName: \"kubernetes.io/projected/f0f41608-5a18-4065-90cb-2c66870cb435-kube-api-access-8g7hs\") on node \"crc\" DevicePath \"\"" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.162874 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f0f41608-5a18-4065-90cb-2c66870cb435-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.699434 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" event={"ID":"f0f41608-5a18-4065-90cb-2c66870cb435","Type":"ContainerDied","Data":"57187ac322569d1d4b36fe886497f108760f9afc90f2413a67a10822a2ed5441"} Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.699481 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57187ac322569d1d4b36fe886497f108760f9afc90f2413a67a10822a2ed5441" Jan 29 15:30:03 crc kubenswrapper[4767]: I0129 15:30:03.699501 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495010-hg69h" Jan 29 15:30:11 crc kubenswrapper[4767]: I0129 15:30:11.018601 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:30:11 crc kubenswrapper[4767]: E0129 15:30:11.019391 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:30:22 crc kubenswrapper[4767]: I0129 15:30:22.018712 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:30:22 crc kubenswrapper[4767]: E0129 15:30:22.019550 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:30:32 crc kubenswrapper[4767]: I0129 15:30:32.951541 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vkwhv"] Jan 29 15:30:32 crc kubenswrapper[4767]: E0129 15:30:32.952389 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0f41608-5a18-4065-90cb-2c66870cb435" containerName="collect-profiles" Jan 29 15:30:32 crc kubenswrapper[4767]: I0129 15:30:32.952403 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0f41608-5a18-4065-90cb-2c66870cb435" containerName="collect-profiles" Jan 29 15:30:32 crc kubenswrapper[4767]: I0129 15:30:32.952544 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0f41608-5a18-4065-90cb-2c66870cb435" containerName="collect-profiles" Jan 29 15:30:32 crc kubenswrapper[4767]: I0129 15:30:32.953604 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:32 crc kubenswrapper[4767]: I0129 15:30:32.966407 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkwhv"] Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.018429 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:30:33 crc kubenswrapper[4767]: E0129 15:30:33.018647 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.038340 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-utilities\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.038417 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x9k8\" (UniqueName: \"kubernetes.io/projected/6d97196d-58bd-4ffd-a7e8-9d76934a654e-kube-api-access-8x9k8\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.038485 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-catalog-content\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.140100 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-utilities\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.140167 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x9k8\" (UniqueName: \"kubernetes.io/projected/6d97196d-58bd-4ffd-a7e8-9d76934a654e-kube-api-access-8x9k8\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.140214 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-catalog-content\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.140729 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-catalog-content\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.140801 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-utilities\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.158361 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x9k8\" (UniqueName: \"kubernetes.io/projected/6d97196d-58bd-4ffd-a7e8-9d76934a654e-kube-api-access-8x9k8\") pod \"certified-operators-vkwhv\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.274880 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.747105 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkwhv"] Jan 29 15:30:33 crc kubenswrapper[4767]: I0129 15:30:33.942329 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwhv" event={"ID":"6d97196d-58bd-4ffd-a7e8-9d76934a654e","Type":"ContainerStarted","Data":"4f93713e5e19599698a7fe743cddc3163eadaafd346609aa74cd8e1190345c25"} Jan 29 15:30:34 crc kubenswrapper[4767]: I0129 15:30:34.950545 4767 generic.go:334] "Generic (PLEG): container finished" podID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerID="11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847" exitCode=0 Jan 29 15:30:34 crc kubenswrapper[4767]: I0129 15:30:34.951113 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwhv" event={"ID":"6d97196d-58bd-4ffd-a7e8-9d76934a654e","Type":"ContainerDied","Data":"11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847"} Jan 29 15:30:37 crc kubenswrapper[4767]: I0129 15:30:37.975436 4767 generic.go:334] "Generic (PLEG): container finished" podID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerID="2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1" exitCode=0 Jan 29 15:30:37 crc kubenswrapper[4767]: I0129 15:30:37.975550 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwhv" event={"ID":"6d97196d-58bd-4ffd-a7e8-9d76934a654e","Type":"ContainerDied","Data":"2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1"} Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.015455 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwhv" event={"ID":"6d97196d-58bd-4ffd-a7e8-9d76934a654e","Type":"ContainerStarted","Data":"b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352"} Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.039881 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vkwhv" podStartSLOduration=4.130806129 podStartE2EDuration="9.039854711s" podCreationTimestamp="2026-01-29 15:30:32 +0000 UTC" firstStartedPulling="2026-01-29 15:30:34.953297922 +0000 UTC m=+1790.763614961" lastFinishedPulling="2026-01-29 15:30:39.862346484 +0000 UTC m=+1795.672663543" observedRunningTime="2026-01-29 15:30:41.03530725 +0000 UTC m=+1796.845624319" watchObservedRunningTime="2026-01-29 15:30:41.039854711 +0000 UTC m=+1796.850171790" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.260802 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vvjd8"] Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.262106 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.276764 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvjd8"] Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.391437 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-catalog-content\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.391769 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9875v\" (UniqueName: \"kubernetes.io/projected/c30032fc-e452-4b23-9b23-9fbe8ed34187-kube-api-access-9875v\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.391869 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-utilities\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.493133 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9875v\" (UniqueName: \"kubernetes.io/projected/c30032fc-e452-4b23-9b23-9fbe8ed34187-kube-api-access-9875v\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.493251 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-utilities\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.493309 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-catalog-content\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.493756 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-utilities\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.493818 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-catalog-content\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.516709 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9875v\" (UniqueName: \"kubernetes.io/projected/c30032fc-e452-4b23-9b23-9fbe8ed34187-kube-api-access-9875v\") pod \"community-operators-vvjd8\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:41 crc kubenswrapper[4767]: I0129 15:30:41.578392 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:42 crc kubenswrapper[4767]: I0129 15:30:42.096360 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvjd8"] Jan 29 15:30:43 crc kubenswrapper[4767]: I0129 15:30:43.031071 4767 generic.go:334] "Generic (PLEG): container finished" podID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerID="188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f" exitCode=0 Jan 29 15:30:43 crc kubenswrapper[4767]: I0129 15:30:43.031112 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvjd8" event={"ID":"c30032fc-e452-4b23-9b23-9fbe8ed34187","Type":"ContainerDied","Data":"188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f"} Jan 29 15:30:43 crc kubenswrapper[4767]: I0129 15:30:43.031135 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvjd8" event={"ID":"c30032fc-e452-4b23-9b23-9fbe8ed34187","Type":"ContainerStarted","Data":"781cc56c7d48c51a31e46e1cf0323e8b574639376fdcb831e58e88f445cb51fc"} Jan 29 15:30:43 crc kubenswrapper[4767]: I0129 15:30:43.276400 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:43 crc kubenswrapper[4767]: I0129 15:30:43.276796 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:43 crc kubenswrapper[4767]: I0129 15:30:43.316054 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:44 crc kubenswrapper[4767]: I0129 15:30:44.019495 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:30:44 crc kubenswrapper[4767]: E0129 15:30:44.019967 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:30:46 crc kubenswrapper[4767]: I0129 15:30:46.057011 4767 generic.go:334] "Generic (PLEG): container finished" podID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerID="56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1" exitCode=0 Jan 29 15:30:46 crc kubenswrapper[4767]: I0129 15:30:46.057067 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvjd8" event={"ID":"c30032fc-e452-4b23-9b23-9fbe8ed34187","Type":"ContainerDied","Data":"56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1"} Jan 29 15:30:50 crc kubenswrapper[4767]: I0129 15:30:50.087479 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvjd8" event={"ID":"c30032fc-e452-4b23-9b23-9fbe8ed34187","Type":"ContainerStarted","Data":"c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6"} Jan 29 15:30:51 crc kubenswrapper[4767]: I0129 15:30:51.118682 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vvjd8" podStartSLOduration=5.533673426 podStartE2EDuration="10.118664299s" podCreationTimestamp="2026-01-29 15:30:41 +0000 UTC" firstStartedPulling="2026-01-29 15:30:43.033400583 +0000 UTC m=+1798.843717632" lastFinishedPulling="2026-01-29 15:30:47.618391466 +0000 UTC m=+1803.428708505" observedRunningTime="2026-01-29 15:30:51.114203922 +0000 UTC m=+1806.924520951" watchObservedRunningTime="2026-01-29 15:30:51.118664299 +0000 UTC m=+1806.928981338" Jan 29 15:30:51 crc kubenswrapper[4767]: I0129 15:30:51.579575 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:51 crc kubenswrapper[4767]: I0129 15:30:51.579965 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:51 crc kubenswrapper[4767]: I0129 15:30:51.620655 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:30:53 crc kubenswrapper[4767]: I0129 15:30:53.324229 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:53 crc kubenswrapper[4767]: I0129 15:30:53.367381 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkwhv"] Jan 29 15:30:54 crc kubenswrapper[4767]: I0129 15:30:54.112208 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vkwhv" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="registry-server" containerID="cri-o://b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352" gracePeriod=2 Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.088088 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.123235 4767 generic.go:334] "Generic (PLEG): container finished" podID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerID="b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352" exitCode=0 Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.123307 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwhv" event={"ID":"6d97196d-58bd-4ffd-a7e8-9d76934a654e","Type":"ContainerDied","Data":"b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352"} Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.123358 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwhv" event={"ID":"6d97196d-58bd-4ffd-a7e8-9d76934a654e","Type":"ContainerDied","Data":"4f93713e5e19599698a7fe743cddc3163eadaafd346609aa74cd8e1190345c25"} Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.123395 4767 scope.go:117] "RemoveContainer" containerID="b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.123622 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwhv" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.159589 4767 scope.go:117] "RemoveContainer" containerID="2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.184058 4767 scope.go:117] "RemoveContainer" containerID="11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.197489 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-utilities\") pod \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.197620 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-catalog-content\") pod \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.197673 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x9k8\" (UniqueName: \"kubernetes.io/projected/6d97196d-58bd-4ffd-a7e8-9d76934a654e-kube-api-access-8x9k8\") pod \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\" (UID: \"6d97196d-58bd-4ffd-a7e8-9d76934a654e\") " Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.199298 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-utilities" (OuterVolumeSpecName: "utilities") pod "6d97196d-58bd-4ffd-a7e8-9d76934a654e" (UID: "6d97196d-58bd-4ffd-a7e8-9d76934a654e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.204316 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d97196d-58bd-4ffd-a7e8-9d76934a654e-kube-api-access-8x9k8" (OuterVolumeSpecName: "kube-api-access-8x9k8") pod "6d97196d-58bd-4ffd-a7e8-9d76934a654e" (UID: "6d97196d-58bd-4ffd-a7e8-9d76934a654e"). InnerVolumeSpecName "kube-api-access-8x9k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.206887 4767 scope.go:117] "RemoveContainer" containerID="b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352" Jan 29 15:30:55 crc kubenswrapper[4767]: E0129 15:30:55.207427 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352\": container with ID starting with b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352 not found: ID does not exist" containerID="b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.207471 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352"} err="failed to get container status \"b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352\": rpc error: code = NotFound desc = could not find container \"b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352\": container with ID starting with b494ecd9e7f78be8b86e6199d3d5d00aed36a5aa6644c97271230eafd47a5352 not found: ID does not exist" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.207499 4767 scope.go:117] "RemoveContainer" containerID="2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1" Jan 29 15:30:55 crc kubenswrapper[4767]: E0129 15:30:55.207823 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1\": container with ID starting with 2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1 not found: ID does not exist" containerID="2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.207879 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1"} err="failed to get container status \"2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1\": rpc error: code = NotFound desc = could not find container \"2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1\": container with ID starting with 2f145c3c9b192685f018073983000acf9a38b2a40a2fb584c23e403c61124ff1 not found: ID does not exist" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.207952 4767 scope.go:117] "RemoveContainer" containerID="11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847" Jan 29 15:30:55 crc kubenswrapper[4767]: E0129 15:30:55.208362 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847\": container with ID starting with 11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847 not found: ID does not exist" containerID="11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.208406 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847"} err="failed to get container status \"11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847\": rpc error: code = NotFound desc = could not find container \"11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847\": container with ID starting with 11df0e1051fe888b64bcd8a175149a3b8df1b5e7a99ee68d6c6acf0332a25847 not found: ID does not exist" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.250473 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d97196d-58bd-4ffd-a7e8-9d76934a654e" (UID: "6d97196d-58bd-4ffd-a7e8-9d76934a654e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.299237 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.299279 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x9k8\" (UniqueName: \"kubernetes.io/projected/6d97196d-58bd-4ffd-a7e8-9d76934a654e-kube-api-access-8x9k8\") on node \"crc\" DevicePath \"\"" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.299288 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d97196d-58bd-4ffd-a7e8-9d76934a654e-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.490150 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkwhv"] Jan 29 15:30:55 crc kubenswrapper[4767]: I0129 15:30:55.497859 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vkwhv"] Jan 29 15:30:57 crc kubenswrapper[4767]: I0129 15:30:57.028600 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" path="/var/lib/kubelet/pods/6d97196d-58bd-4ffd-a7e8-9d76934a654e/volumes" Jan 29 15:30:59 crc kubenswrapper[4767]: I0129 15:30:59.019138 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:30:59 crc kubenswrapper[4767]: E0129 15:30:59.019619 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:31:01 crc kubenswrapper[4767]: I0129 15:31:01.623338 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:31:01 crc kubenswrapper[4767]: I0129 15:31:01.671794 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvjd8"] Jan 29 15:31:02 crc kubenswrapper[4767]: I0129 15:31:02.174305 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vvjd8" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="registry-server" containerID="cri-o://c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6" gracePeriod=2 Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.140768 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.184958 4767 generic.go:334] "Generic (PLEG): container finished" podID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerID="c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6" exitCode=0 Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.185000 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvjd8" event={"ID":"c30032fc-e452-4b23-9b23-9fbe8ed34187","Type":"ContainerDied","Data":"c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6"} Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.185016 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvjd8" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.185036 4767 scope.go:117] "RemoveContainer" containerID="c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.185025 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvjd8" event={"ID":"c30032fc-e452-4b23-9b23-9fbe8ed34187","Type":"ContainerDied","Data":"781cc56c7d48c51a31e46e1cf0323e8b574639376fdcb831e58e88f445cb51fc"} Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.205120 4767 scope.go:117] "RemoveContainer" containerID="56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.217261 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-catalog-content\") pod \"c30032fc-e452-4b23-9b23-9fbe8ed34187\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.217369 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9875v\" (UniqueName: \"kubernetes.io/projected/c30032fc-e452-4b23-9b23-9fbe8ed34187-kube-api-access-9875v\") pod \"c30032fc-e452-4b23-9b23-9fbe8ed34187\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.217400 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-utilities\") pod \"c30032fc-e452-4b23-9b23-9fbe8ed34187\" (UID: \"c30032fc-e452-4b23-9b23-9fbe8ed34187\") " Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.218507 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-utilities" (OuterVolumeSpecName: "utilities") pod "c30032fc-e452-4b23-9b23-9fbe8ed34187" (UID: "c30032fc-e452-4b23-9b23-9fbe8ed34187"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.222953 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c30032fc-e452-4b23-9b23-9fbe8ed34187-kube-api-access-9875v" (OuterVolumeSpecName: "kube-api-access-9875v") pod "c30032fc-e452-4b23-9b23-9fbe8ed34187" (UID: "c30032fc-e452-4b23-9b23-9fbe8ed34187"). InnerVolumeSpecName "kube-api-access-9875v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.223263 4767 scope.go:117] "RemoveContainer" containerID="188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.270457 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c30032fc-e452-4b23-9b23-9fbe8ed34187" (UID: "c30032fc-e452-4b23-9b23-9fbe8ed34187"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.276159 4767 scope.go:117] "RemoveContainer" containerID="c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6" Jan 29 15:31:03 crc kubenswrapper[4767]: E0129 15:31:03.276625 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6\": container with ID starting with c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6 not found: ID does not exist" containerID="c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.276659 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6"} err="failed to get container status \"c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6\": rpc error: code = NotFound desc = could not find container \"c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6\": container with ID starting with c50abf5b514285cf00ca774e1602f5ae7bb51da2650d25d57a9005296434dee6 not found: ID does not exist" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.276683 4767 scope.go:117] "RemoveContainer" containerID="56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1" Jan 29 15:31:03 crc kubenswrapper[4767]: E0129 15:31:03.277500 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1\": container with ID starting with 56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1 not found: ID does not exist" containerID="56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.277543 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1"} err="failed to get container status \"56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1\": rpc error: code = NotFound desc = could not find container \"56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1\": container with ID starting with 56f252f3d2fd30d53fa71a493fb93fa16dfc92491b52dc5a820809ecc5cf89e1 not found: ID does not exist" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.277575 4767 scope.go:117] "RemoveContainer" containerID="188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f" Jan 29 15:31:03 crc kubenswrapper[4767]: E0129 15:31:03.278069 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f\": container with ID starting with 188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f not found: ID does not exist" containerID="188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.278093 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f"} err="failed to get container status \"188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f\": rpc error: code = NotFound desc = could not find container \"188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f\": container with ID starting with 188d5586222d433fef68c27fff47b159849ce0b312c8cfbf6d8cbc1b41a8747f not found: ID does not exist" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.319409 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.319767 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9875v\" (UniqueName: \"kubernetes.io/projected/c30032fc-e452-4b23-9b23-9fbe8ed34187-kube-api-access-9875v\") on node \"crc\" DevicePath \"\"" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.319779 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c30032fc-e452-4b23-9b23-9fbe8ed34187-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.523178 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvjd8"] Jan 29 15:31:03 crc kubenswrapper[4767]: I0129 15:31:03.531054 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vvjd8"] Jan 29 15:31:05 crc kubenswrapper[4767]: I0129 15:31:05.031230 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" path="/var/lib/kubelet/pods/c30032fc-e452-4b23-9b23-9fbe8ed34187/volumes" Jan 29 15:31:11 crc kubenswrapper[4767]: I0129 15:31:11.018786 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:31:11 crc kubenswrapper[4767]: E0129 15:31:11.019973 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:31:26 crc kubenswrapper[4767]: I0129 15:31:26.018542 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:31:26 crc kubenswrapper[4767]: E0129 15:31:26.019366 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:31:40 crc kubenswrapper[4767]: I0129 15:31:40.018479 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:31:40 crc kubenswrapper[4767]: E0129 15:31:40.019666 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:31:53 crc kubenswrapper[4767]: I0129 15:31:53.019142 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:31:53 crc kubenswrapper[4767]: E0129 15:31:53.019953 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:32:07 crc kubenswrapper[4767]: I0129 15:32:07.019019 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:32:07 crc kubenswrapper[4767]: E0129 15:32:07.019795 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:32:20 crc kubenswrapper[4767]: I0129 15:32:20.018711 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:32:20 crc kubenswrapper[4767]: E0129 15:32:20.019428 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:32:35 crc kubenswrapper[4767]: I0129 15:32:35.022651 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:32:35 crc kubenswrapper[4767]: E0129 15:32:35.024044 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:32:48 crc kubenswrapper[4767]: I0129 15:32:48.019354 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:32:48 crc kubenswrapper[4767]: E0129 15:32:48.020158 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:33:01 crc kubenswrapper[4767]: I0129 15:33:01.019034 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:33:01 crc kubenswrapper[4767]: E0129 15:33:01.019801 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:33:14 crc kubenswrapper[4767]: I0129 15:33:14.019073 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:33:14 crc kubenswrapper[4767]: E0129 15:33:14.019800 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:33:26 crc kubenswrapper[4767]: I0129 15:33:26.020048 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:33:26 crc kubenswrapper[4767]: E0129 15:33:26.021827 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:33:38 crc kubenswrapper[4767]: I0129 15:33:38.018399 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:33:38 crc kubenswrapper[4767]: E0129 15:33:38.019194 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:33:49 crc kubenswrapper[4767]: I0129 15:33:49.018484 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:33:49 crc kubenswrapper[4767]: I0129 15:33:49.680809 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"a76200482fc057bb3597e2c677ad5fb7b0b8c2f618249492d21ac03db19a14c6"} Jan 29 15:36:12 crc kubenswrapper[4767]: I0129 15:36:12.805916 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:36:12 crc kubenswrapper[4767]: I0129 15:36:12.806420 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:36:42 crc kubenswrapper[4767]: I0129 15:36:42.805374 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:36:42 crc kubenswrapper[4767]: I0129 15:36:42.805803 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:37:12 crc kubenswrapper[4767]: I0129 15:37:12.805919 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:37:12 crc kubenswrapper[4767]: I0129 15:37:12.806918 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:37:12 crc kubenswrapper[4767]: I0129 15:37:12.806992 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:37:12 crc kubenswrapper[4767]: I0129 15:37:12.807974 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a76200482fc057bb3597e2c677ad5fb7b0b8c2f618249492d21ac03db19a14c6"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:37:12 crc kubenswrapper[4767]: I0129 15:37:12.808046 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://a76200482fc057bb3597e2c677ad5fb7b0b8c2f618249492d21ac03db19a14c6" gracePeriod=600 Jan 29 15:37:13 crc kubenswrapper[4767]: I0129 15:37:13.311339 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="a76200482fc057bb3597e2c677ad5fb7b0b8c2f618249492d21ac03db19a14c6" exitCode=0 Jan 29 15:37:13 crc kubenswrapper[4767]: I0129 15:37:13.311388 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"a76200482fc057bb3597e2c677ad5fb7b0b8c2f618249492d21ac03db19a14c6"} Jan 29 15:37:13 crc kubenswrapper[4767]: I0129 15:37:13.311885 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4"} Jan 29 15:37:13 crc kubenswrapper[4767]: I0129 15:37:13.311932 4767 scope.go:117] "RemoveContainer" containerID="8121cf25bd45deeb18e86eecc113284ac03b892aa7f380ff257f8a7df30749ce" Jan 29 15:39:42 crc kubenswrapper[4767]: I0129 15:39:42.805231 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:39:42 crc kubenswrapper[4767]: I0129 15:39:42.805783 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.422551 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r5t49"] Jan 29 15:39:43 crc kubenswrapper[4767]: E0129 15:39:43.422953 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="extract-content" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.422972 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="extract-content" Jan 29 15:39:43 crc kubenswrapper[4767]: E0129 15:39:43.422994 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="extract-content" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.423006 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="extract-content" Jan 29 15:39:43 crc kubenswrapper[4767]: E0129 15:39:43.423023 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="registry-server" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.423037 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="registry-server" Jan 29 15:39:43 crc kubenswrapper[4767]: E0129 15:39:43.423068 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="extract-utilities" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.423081 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="extract-utilities" Jan 29 15:39:43 crc kubenswrapper[4767]: E0129 15:39:43.423106 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="extract-utilities" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.423118 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="extract-utilities" Jan 29 15:39:43 crc kubenswrapper[4767]: E0129 15:39:43.423140 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="registry-server" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.423152 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="registry-server" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.423385 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="c30032fc-e452-4b23-9b23-9fbe8ed34187" containerName="registry-server" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.423407 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d97196d-58bd-4ffd-a7e8-9d76934a654e" containerName="registry-server" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.425045 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.477695 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5t49"] Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.540809 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-catalog-content\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.540935 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sfwm\" (UniqueName: \"kubernetes.io/projected/6e04a22c-8d63-4a94-9417-633b5ea47737-kube-api-access-6sfwm\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.541009 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-utilities\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.642701 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-utilities\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.642789 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-catalog-content\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.642866 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sfwm\" (UniqueName: \"kubernetes.io/projected/6e04a22c-8d63-4a94-9417-633b5ea47737-kube-api-access-6sfwm\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.643314 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-utilities\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.643347 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-catalog-content\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.663190 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sfwm\" (UniqueName: \"kubernetes.io/projected/6e04a22c-8d63-4a94-9417-633b5ea47737-kube-api-access-6sfwm\") pod \"redhat-operators-r5t49\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:43 crc kubenswrapper[4767]: I0129 15:39:43.756613 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:44 crc kubenswrapper[4767]: I0129 15:39:44.029739 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5t49"] Jan 29 15:39:44 crc kubenswrapper[4767]: I0129 15:39:44.780120 4767 generic.go:334] "Generic (PLEG): container finished" podID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerID="549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0" exitCode=0 Jan 29 15:39:44 crc kubenswrapper[4767]: I0129 15:39:44.780216 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5t49" event={"ID":"6e04a22c-8d63-4a94-9417-633b5ea47737","Type":"ContainerDied","Data":"549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0"} Jan 29 15:39:44 crc kubenswrapper[4767]: I0129 15:39:44.780442 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5t49" event={"ID":"6e04a22c-8d63-4a94-9417-633b5ea47737","Type":"ContainerStarted","Data":"85b6765117540f2463c271f6c811fa07bff94433439f4e85f1c0da9371ca3b4b"} Jan 29 15:39:44 crc kubenswrapper[4767]: I0129 15:39:44.781655 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:39:45 crc kubenswrapper[4767]: I0129 15:39:45.793861 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5t49" event={"ID":"6e04a22c-8d63-4a94-9417-633b5ea47737","Type":"ContainerStarted","Data":"eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc"} Jan 29 15:39:46 crc kubenswrapper[4767]: I0129 15:39:46.806124 4767 generic.go:334] "Generic (PLEG): container finished" podID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerID="eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc" exitCode=0 Jan 29 15:39:46 crc kubenswrapper[4767]: I0129 15:39:46.806176 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5t49" event={"ID":"6e04a22c-8d63-4a94-9417-633b5ea47737","Type":"ContainerDied","Data":"eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc"} Jan 29 15:39:47 crc kubenswrapper[4767]: I0129 15:39:47.815584 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5t49" event={"ID":"6e04a22c-8d63-4a94-9417-633b5ea47737","Type":"ContainerStarted","Data":"4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc"} Jan 29 15:39:47 crc kubenswrapper[4767]: I0129 15:39:47.839654 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r5t49" podStartSLOduration=2.376713877 podStartE2EDuration="4.839638211s" podCreationTimestamp="2026-01-29 15:39:43 +0000 UTC" firstStartedPulling="2026-01-29 15:39:44.781363345 +0000 UTC m=+2340.591680384" lastFinishedPulling="2026-01-29 15:39:47.244287629 +0000 UTC m=+2343.054604718" observedRunningTime="2026-01-29 15:39:47.838030776 +0000 UTC m=+2343.648347825" watchObservedRunningTime="2026-01-29 15:39:47.839638211 +0000 UTC m=+2343.649955250" Jan 29 15:39:53 crc kubenswrapper[4767]: I0129 15:39:53.757802 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:53 crc kubenswrapper[4767]: I0129 15:39:53.758652 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:39:54 crc kubenswrapper[4767]: I0129 15:39:54.827403 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-r5t49" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="registry-server" probeResult="failure" output=< Jan 29 15:39:54 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:39:54 crc kubenswrapper[4767]: > Jan 29 15:40:03 crc kubenswrapper[4767]: I0129 15:40:03.798745 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:40:03 crc kubenswrapper[4767]: I0129 15:40:03.849644 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:40:04 crc kubenswrapper[4767]: I0129 15:40:04.044123 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5t49"] Jan 29 15:40:04 crc kubenswrapper[4767]: I0129 15:40:04.944070 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r5t49" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="registry-server" containerID="cri-o://4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc" gracePeriod=2 Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.829325 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.849196 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-utilities\") pod \"6e04a22c-8d63-4a94-9417-633b5ea47737\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.849351 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sfwm\" (UniqueName: \"kubernetes.io/projected/6e04a22c-8d63-4a94-9417-633b5ea47737-kube-api-access-6sfwm\") pod \"6e04a22c-8d63-4a94-9417-633b5ea47737\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.849440 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-catalog-content\") pod \"6e04a22c-8d63-4a94-9417-633b5ea47737\" (UID: \"6e04a22c-8d63-4a94-9417-633b5ea47737\") " Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.853813 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-utilities" (OuterVolumeSpecName: "utilities") pod "6e04a22c-8d63-4a94-9417-633b5ea47737" (UID: "6e04a22c-8d63-4a94-9417-633b5ea47737"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.858869 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e04a22c-8d63-4a94-9417-633b5ea47737-kube-api-access-6sfwm" (OuterVolumeSpecName: "kube-api-access-6sfwm") pod "6e04a22c-8d63-4a94-9417-633b5ea47737" (UID: "6e04a22c-8d63-4a94-9417-633b5ea47737"). InnerVolumeSpecName "kube-api-access-6sfwm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.954325 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.954358 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sfwm\" (UniqueName: \"kubernetes.io/projected/6e04a22c-8d63-4a94-9417-633b5ea47737-kube-api-access-6sfwm\") on node \"crc\" DevicePath \"\"" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.958810 4767 generic.go:334] "Generic (PLEG): container finished" podID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerID="4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc" exitCode=0 Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.958849 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5t49" event={"ID":"6e04a22c-8d63-4a94-9417-633b5ea47737","Type":"ContainerDied","Data":"4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc"} Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.958875 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5t49" event={"ID":"6e04a22c-8d63-4a94-9417-633b5ea47737","Type":"ContainerDied","Data":"85b6765117540f2463c271f6c811fa07bff94433439f4e85f1c0da9371ca3b4b"} Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.958908 4767 scope.go:117] "RemoveContainer" containerID="4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.959261 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5t49" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.977103 4767 scope.go:117] "RemoveContainer" containerID="eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc" Jan 29 15:40:05 crc kubenswrapper[4767]: I0129 15:40:05.992086 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e04a22c-8d63-4a94-9417-633b5ea47737" (UID: "6e04a22c-8d63-4a94-9417-633b5ea47737"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.008876 4767 scope.go:117] "RemoveContainer" containerID="549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.028388 4767 scope.go:117] "RemoveContainer" containerID="4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc" Jan 29 15:40:06 crc kubenswrapper[4767]: E0129 15:40:06.028848 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc\": container with ID starting with 4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc not found: ID does not exist" containerID="4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.029124 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc"} err="failed to get container status \"4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc\": rpc error: code = NotFound desc = could not find container \"4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc\": container with ID starting with 4f7c3bcb8c906700c390af0cd018b6592cf435a40ed2b4e83e7883f0f94817bc not found: ID does not exist" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.029267 4767 scope.go:117] "RemoveContainer" containerID="eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc" Jan 29 15:40:06 crc kubenswrapper[4767]: E0129 15:40:06.029783 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc\": container with ID starting with eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc not found: ID does not exist" containerID="eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.029819 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc"} err="failed to get container status \"eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc\": rpc error: code = NotFound desc = could not find container \"eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc\": container with ID starting with eaa718b6a999fbe8397ab55b9e228d206f2bf3959db33554504b3c60fee0d4dc not found: ID does not exist" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.029839 4767 scope.go:117] "RemoveContainer" containerID="549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0" Jan 29 15:40:06 crc kubenswrapper[4767]: E0129 15:40:06.030464 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0\": container with ID starting with 549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0 not found: ID does not exist" containerID="549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.030498 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0"} err="failed to get container status \"549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0\": rpc error: code = NotFound desc = could not find container \"549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0\": container with ID starting with 549bca7aaa03e4fc735ae0070ea777d660cf2164f274dbc727f2c92b697e8dd0 not found: ID does not exist" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.056572 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e04a22c-8d63-4a94-9417-633b5ea47737-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.294247 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5t49"] Jan 29 15:40:06 crc kubenswrapper[4767]: I0129 15:40:06.301140 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r5t49"] Jan 29 15:40:07 crc kubenswrapper[4767]: I0129 15:40:07.026482 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" path="/var/lib/kubelet/pods/6e04a22c-8d63-4a94-9417-633b5ea47737/volumes" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.248528 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-86fv7"] Jan 29 15:40:09 crc kubenswrapper[4767]: E0129 15:40:09.249849 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="extract-content" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.249878 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="extract-content" Jan 29 15:40:09 crc kubenswrapper[4767]: E0129 15:40:09.249933 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="registry-server" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.249953 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="registry-server" Jan 29 15:40:09 crc kubenswrapper[4767]: E0129 15:40:09.249975 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="extract-utilities" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.249995 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="extract-utilities" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.250400 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e04a22c-8d63-4a94-9417-633b5ea47737" containerName="registry-server" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.253015 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.265262 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86fv7"] Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.403936 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpmxk\" (UniqueName: \"kubernetes.io/projected/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-kube-api-access-mpmxk\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.404019 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-catalog-content\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.404136 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-utilities\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.506449 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-utilities\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.507015 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpmxk\" (UniqueName: \"kubernetes.io/projected/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-kube-api-access-mpmxk\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.507043 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-catalog-content\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.507299 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-catalog-content\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.506921 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-utilities\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.529930 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpmxk\" (UniqueName: \"kubernetes.io/projected/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-kube-api-access-mpmxk\") pod \"redhat-marketplace-86fv7\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:09 crc kubenswrapper[4767]: I0129 15:40:09.588211 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:10 crc kubenswrapper[4767]: I0129 15:40:10.059902 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-86fv7"] Jan 29 15:40:10 crc kubenswrapper[4767]: I0129 15:40:10.990633 4767 generic.go:334] "Generic (PLEG): container finished" podID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerID="e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e" exitCode=0 Jan 29 15:40:10 crc kubenswrapper[4767]: I0129 15:40:10.990741 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86fv7" event={"ID":"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4","Type":"ContainerDied","Data":"e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e"} Jan 29 15:40:10 crc kubenswrapper[4767]: I0129 15:40:10.991020 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86fv7" event={"ID":"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4","Type":"ContainerStarted","Data":"3f087bba0b444f6b12e533d9dc22b00b7f71ff4aa74758c0887f565c9ef82f8b"} Jan 29 15:40:12 crc kubenswrapper[4767]: I0129 15:40:12.805641 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:40:12 crc kubenswrapper[4767]: I0129 15:40:12.805704 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:40:13 crc kubenswrapper[4767]: I0129 15:40:13.008125 4767 generic.go:334] "Generic (PLEG): container finished" podID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerID="24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116" exitCode=0 Jan 29 15:40:13 crc kubenswrapper[4767]: I0129 15:40:13.008196 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86fv7" event={"ID":"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4","Type":"ContainerDied","Data":"24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116"} Jan 29 15:40:14 crc kubenswrapper[4767]: I0129 15:40:14.018487 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86fv7" event={"ID":"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4","Type":"ContainerStarted","Data":"c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd"} Jan 29 15:40:14 crc kubenswrapper[4767]: I0129 15:40:14.036699 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-86fv7" podStartSLOduration=2.319794819 podStartE2EDuration="5.036679633s" podCreationTimestamp="2026-01-29 15:40:09 +0000 UTC" firstStartedPulling="2026-01-29 15:40:10.993546515 +0000 UTC m=+2366.803863554" lastFinishedPulling="2026-01-29 15:40:13.710431329 +0000 UTC m=+2369.520748368" observedRunningTime="2026-01-29 15:40:14.035793058 +0000 UTC m=+2369.846110127" watchObservedRunningTime="2026-01-29 15:40:14.036679633 +0000 UTC m=+2369.846996692" Jan 29 15:40:19 crc kubenswrapper[4767]: I0129 15:40:19.589257 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:19 crc kubenswrapper[4767]: I0129 15:40:19.590878 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:19 crc kubenswrapper[4767]: I0129 15:40:19.636880 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:20 crc kubenswrapper[4767]: I0129 15:40:20.110751 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:20 crc kubenswrapper[4767]: I0129 15:40:20.161123 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86fv7"] Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.078204 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-86fv7" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="registry-server" containerID="cri-o://c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd" gracePeriod=2 Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.450536 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.520112 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-catalog-content\") pod \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.520226 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpmxk\" (UniqueName: \"kubernetes.io/projected/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-kube-api-access-mpmxk\") pod \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.520248 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-utilities\") pod \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\" (UID: \"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4\") " Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.521363 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-utilities" (OuterVolumeSpecName: "utilities") pod "d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" (UID: "d82b6e64-13ff-41b2-b752-a06a6dd5b7e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.526463 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-kube-api-access-mpmxk" (OuterVolumeSpecName: "kube-api-access-mpmxk") pod "d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" (UID: "d82b6e64-13ff-41b2-b752-a06a6dd5b7e4"). InnerVolumeSpecName "kube-api-access-mpmxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.538710 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" (UID: "d82b6e64-13ff-41b2-b752-a06a6dd5b7e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.622177 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.622220 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpmxk\" (UniqueName: \"kubernetes.io/projected/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-kube-api-access-mpmxk\") on node \"crc\" DevicePath \"\"" Jan 29 15:40:22 crc kubenswrapper[4767]: I0129 15:40:22.622235 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.087872 4767 generic.go:334] "Generic (PLEG): container finished" podID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerID="c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd" exitCode=0 Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.087949 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-86fv7" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.087946 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86fv7" event={"ID":"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4","Type":"ContainerDied","Data":"c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd"} Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.088155 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-86fv7" event={"ID":"d82b6e64-13ff-41b2-b752-a06a6dd5b7e4","Type":"ContainerDied","Data":"3f087bba0b444f6b12e533d9dc22b00b7f71ff4aa74758c0887f565c9ef82f8b"} Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.088182 4767 scope.go:117] "RemoveContainer" containerID="c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.109051 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-86fv7"] Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.113151 4767 scope.go:117] "RemoveContainer" containerID="24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.119189 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-86fv7"] Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.141867 4767 scope.go:117] "RemoveContainer" containerID="e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.158795 4767 scope.go:117] "RemoveContainer" containerID="c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd" Jan 29 15:40:23 crc kubenswrapper[4767]: E0129 15:40:23.159333 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd\": container with ID starting with c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd not found: ID does not exist" containerID="c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.159380 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd"} err="failed to get container status \"c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd\": rpc error: code = NotFound desc = could not find container \"c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd\": container with ID starting with c2a3f37ca41d1d13ab61a1068ccc53bb6dee9f74d438c9f29ec0428bd49359fd not found: ID does not exist" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.159405 4767 scope.go:117] "RemoveContainer" containerID="24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116" Jan 29 15:40:23 crc kubenswrapper[4767]: E0129 15:40:23.159722 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116\": container with ID starting with 24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116 not found: ID does not exist" containerID="24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.159762 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116"} err="failed to get container status \"24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116\": rpc error: code = NotFound desc = could not find container \"24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116\": container with ID starting with 24e9a99fe4185302c3deb86f4c03f3155d741d743373da0cb990390845595116 not found: ID does not exist" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.159789 4767 scope.go:117] "RemoveContainer" containerID="e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e" Jan 29 15:40:23 crc kubenswrapper[4767]: E0129 15:40:23.160032 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e\": container with ID starting with e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e not found: ID does not exist" containerID="e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e" Jan 29 15:40:23 crc kubenswrapper[4767]: I0129 15:40:23.160060 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e"} err="failed to get container status \"e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e\": rpc error: code = NotFound desc = could not find container \"e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e\": container with ID starting with e76ab894984e5a487fb541af1d6cefe5c19d0532f550e486801ad571b6cefa0e not found: ID does not exist" Jan 29 15:40:25 crc kubenswrapper[4767]: I0129 15:40:25.028976 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" path="/var/lib/kubelet/pods/d82b6e64-13ff-41b2-b752-a06a6dd5b7e4/volumes" Jan 29 15:40:42 crc kubenswrapper[4767]: I0129 15:40:42.805938 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:40:42 crc kubenswrapper[4767]: I0129 15:40:42.807613 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:40:42 crc kubenswrapper[4767]: I0129 15:40:42.807776 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:40:42 crc kubenswrapper[4767]: I0129 15:40:42.808511 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:40:42 crc kubenswrapper[4767]: I0129 15:40:42.808649 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" gracePeriod=600 Jan 29 15:40:42 crc kubenswrapper[4767]: E0129 15:40:42.944160 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:40:43 crc kubenswrapper[4767]: I0129 15:40:43.226344 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" exitCode=0 Jan 29 15:40:43 crc kubenswrapper[4767]: I0129 15:40:43.226386 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4"} Jan 29 15:40:43 crc kubenswrapper[4767]: I0129 15:40:43.226417 4767 scope.go:117] "RemoveContainer" containerID="a76200482fc057bb3597e2c677ad5fb7b0b8c2f618249492d21ac03db19a14c6" Jan 29 15:40:43 crc kubenswrapper[4767]: I0129 15:40:43.226891 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:40:43 crc kubenswrapper[4767]: E0129 15:40:43.227332 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:40:58 crc kubenswrapper[4767]: I0129 15:40:58.020006 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:40:58 crc kubenswrapper[4767]: E0129 15:40:58.020845 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:41:09 crc kubenswrapper[4767]: I0129 15:41:09.018731 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:41:09 crc kubenswrapper[4767]: E0129 15:41:09.019463 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:41:21 crc kubenswrapper[4767]: I0129 15:41:21.018620 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:41:21 crc kubenswrapper[4767]: E0129 15:41:21.019486 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:41:32 crc kubenswrapper[4767]: I0129 15:41:32.018846 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:41:32 crc kubenswrapper[4767]: E0129 15:41:32.019885 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:41:45 crc kubenswrapper[4767]: I0129 15:41:45.022022 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:41:45 crc kubenswrapper[4767]: E0129 15:41:45.022705 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.617787 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xsqfw"] Jan 29 15:41:47 crc kubenswrapper[4767]: E0129 15:41:47.618123 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="extract-utilities" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.618140 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="extract-utilities" Jan 29 15:41:47 crc kubenswrapper[4767]: E0129 15:41:47.618161 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="extract-content" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.618168 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="extract-content" Jan 29 15:41:47 crc kubenswrapper[4767]: E0129 15:41:47.618187 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="registry-server" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.618198 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="registry-server" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.618390 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d82b6e64-13ff-41b2-b752-a06a6dd5b7e4" containerName="registry-server" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.619556 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.624735 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xsqfw"] Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.724515 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-catalog-content\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.724932 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-utilities\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.724970 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75pk6\" (UniqueName: \"kubernetes.io/projected/405e1951-f911-428a-ad5f-2f1f29d76506-kube-api-access-75pk6\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.825812 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-catalog-content\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.825945 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-utilities\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.825990 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75pk6\" (UniqueName: \"kubernetes.io/projected/405e1951-f911-428a-ad5f-2f1f29d76506-kube-api-access-75pk6\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.826494 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-catalog-content\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.826525 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-utilities\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.848352 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75pk6\" (UniqueName: \"kubernetes.io/projected/405e1951-f911-428a-ad5f-2f1f29d76506-kube-api-access-75pk6\") pod \"community-operators-xsqfw\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:47 crc kubenswrapper[4767]: I0129 15:41:47.941405 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:41:48 crc kubenswrapper[4767]: I0129 15:41:48.490179 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xsqfw"] Jan 29 15:41:48 crc kubenswrapper[4767]: I0129 15:41:48.807404 4767 generic.go:334] "Generic (PLEG): container finished" podID="405e1951-f911-428a-ad5f-2f1f29d76506" containerID="93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75" exitCode=0 Jan 29 15:41:48 crc kubenswrapper[4767]: I0129 15:41:48.807484 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsqfw" event={"ID":"405e1951-f911-428a-ad5f-2f1f29d76506","Type":"ContainerDied","Data":"93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75"} Jan 29 15:41:48 crc kubenswrapper[4767]: I0129 15:41:48.807634 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsqfw" event={"ID":"405e1951-f911-428a-ad5f-2f1f29d76506","Type":"ContainerStarted","Data":"ce00a99434da761e3ea1459914058e3a88285d1f730d61f548f5d7074a3d0404"} Jan 29 15:41:48 crc kubenswrapper[4767]: E0129 15:41:48.947775 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:41:48 crc kubenswrapper[4767]: E0129 15:41:48.948186 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75pk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsqfw_openshift-marketplace(405e1951-f911-428a-ad5f-2f1f29d76506): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:41:48 crc kubenswrapper[4767]: E0129 15:41:48.949990 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:41:49 crc kubenswrapper[4767]: E0129 15:41:49.817826 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:42:00 crc kubenswrapper[4767]: I0129 15:42:00.018967 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:42:00 crc kubenswrapper[4767]: E0129 15:42:00.019797 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:42:02 crc kubenswrapper[4767]: E0129 15:42:02.169853 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:42:02 crc kubenswrapper[4767]: E0129 15:42:02.170098 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75pk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsqfw_openshift-marketplace(405e1951-f911-428a-ad5f-2f1f29d76506): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:42:02 crc kubenswrapper[4767]: E0129 15:42:02.171269 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:42:13 crc kubenswrapper[4767]: I0129 15:42:13.019126 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:42:13 crc kubenswrapper[4767]: E0129 15:42:13.019821 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:42:14 crc kubenswrapper[4767]: E0129 15:42:14.022922 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:42:26 crc kubenswrapper[4767]: I0129 15:42:26.019180 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:42:26 crc kubenswrapper[4767]: E0129 15:42:26.020234 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:42:28 crc kubenswrapper[4767]: E0129 15:42:28.149048 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:42:28 crc kubenswrapper[4767]: E0129 15:42:28.149554 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75pk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsqfw_openshift-marketplace(405e1951-f911-428a-ad5f-2f1f29d76506): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:42:28 crc kubenswrapper[4767]: E0129 15:42:28.150822 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:42:41 crc kubenswrapper[4767]: I0129 15:42:41.019556 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:42:41 crc kubenswrapper[4767]: E0129 15:42:41.020422 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:42:43 crc kubenswrapper[4767]: E0129 15:42:43.021451 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:42:55 crc kubenswrapper[4767]: I0129 15:42:55.027225 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:42:55 crc kubenswrapper[4767]: E0129 15:42:55.028459 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:42:58 crc kubenswrapper[4767]: E0129 15:42:58.020985 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:43:09 crc kubenswrapper[4767]: I0129 15:43:09.018404 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:43:09 crc kubenswrapper[4767]: E0129 15:43:09.019224 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:43:13 crc kubenswrapper[4767]: E0129 15:43:13.166302 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:43:13 crc kubenswrapper[4767]: E0129 15:43:13.166785 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75pk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsqfw_openshift-marketplace(405e1951-f911-428a-ad5f-2f1f29d76506): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:43:13 crc kubenswrapper[4767]: E0129 15:43:13.167956 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:43:23 crc kubenswrapper[4767]: I0129 15:43:23.019241 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:43:23 crc kubenswrapper[4767]: E0129 15:43:23.020045 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:43:29 crc kubenswrapper[4767]: E0129 15:43:29.021027 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:43:36 crc kubenswrapper[4767]: I0129 15:43:36.019454 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:43:36 crc kubenswrapper[4767]: E0129 15:43:36.020375 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:43:42 crc kubenswrapper[4767]: E0129 15:43:42.020170 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:43:50 crc kubenswrapper[4767]: I0129 15:43:50.019252 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:43:50 crc kubenswrapper[4767]: E0129 15:43:50.021178 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:43:53 crc kubenswrapper[4767]: E0129 15:43:53.021242 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:44:01 crc kubenswrapper[4767]: I0129 15:44:01.018713 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:44:01 crc kubenswrapper[4767]: E0129 15:44:01.019617 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:44:04 crc kubenswrapper[4767]: E0129 15:44:04.020619 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:44:14 crc kubenswrapper[4767]: I0129 15:44:14.018379 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:44:14 crc kubenswrapper[4767]: E0129 15:44:14.019245 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:44:18 crc kubenswrapper[4767]: E0129 15:44:18.021444 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:44:26 crc kubenswrapper[4767]: I0129 15:44:26.019480 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:44:26 crc kubenswrapper[4767]: E0129 15:44:26.020476 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:44:33 crc kubenswrapper[4767]: E0129 15:44:33.021538 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:44:41 crc kubenswrapper[4767]: I0129 15:44:41.018480 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:44:41 crc kubenswrapper[4767]: E0129 15:44:41.019187 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:44:45 crc kubenswrapper[4767]: I0129 15:44:45.040661 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:44:45 crc kubenswrapper[4767]: E0129 15:44:45.312098 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:44:45 crc kubenswrapper[4767]: E0129 15:44:45.312254 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75pk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsqfw_openshift-marketplace(405e1951-f911-428a-ad5f-2f1f29d76506): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:44:45 crc kubenswrapper[4767]: E0129 15:44:45.313444 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:44:53 crc kubenswrapper[4767]: I0129 15:44:53.687352 4767 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-qcf7k" podUID="1e14bfd9-9551-45e2-85cd-75f27bc10ada" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.53:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 29 15:44:56 crc kubenswrapper[4767]: I0129 15:44:56.019241 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:44:56 crc kubenswrapper[4767]: E0129 15:44:56.020054 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:44:56 crc kubenswrapper[4767]: E0129 15:44:56.021367 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.157066 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n"] Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.161424 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.164628 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.165201 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.166282 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n"] Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.341386 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bac5b3d-af3b-489d-a798-10e2e5d6e221-secret-volume\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.341698 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bac5b3d-af3b-489d-a798-10e2e5d6e221-config-volume\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.341797 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5tpw\" (UniqueName: \"kubernetes.io/projected/1bac5b3d-af3b-489d-a798-10e2e5d6e221-kube-api-access-q5tpw\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.443155 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bac5b3d-af3b-489d-a798-10e2e5d6e221-secret-volume\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.443283 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bac5b3d-af3b-489d-a798-10e2e5d6e221-config-volume\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.443319 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5tpw\" (UniqueName: \"kubernetes.io/projected/1bac5b3d-af3b-489d-a798-10e2e5d6e221-kube-api-access-q5tpw\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.444367 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bac5b3d-af3b-489d-a798-10e2e5d6e221-config-volume\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.450775 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bac5b3d-af3b-489d-a798-10e2e5d6e221-secret-volume\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.462590 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5tpw\" (UniqueName: \"kubernetes.io/projected/1bac5b3d-af3b-489d-a798-10e2e5d6e221-kube-api-access-q5tpw\") pod \"collect-profiles-29495025-h4v8n\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.494154 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:00 crc kubenswrapper[4767]: I0129 15:45:00.936862 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n"] Jan 29 15:45:01 crc kubenswrapper[4767]: I0129 15:45:01.717643 4767 generic.go:334] "Generic (PLEG): container finished" podID="1bac5b3d-af3b-489d-a798-10e2e5d6e221" containerID="67cb80f394a3833393947acb8d504f686f136932499fbad8df5b0076586d2ccc" exitCode=0 Jan 29 15:45:01 crc kubenswrapper[4767]: I0129 15:45:01.717857 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" event={"ID":"1bac5b3d-af3b-489d-a798-10e2e5d6e221","Type":"ContainerDied","Data":"67cb80f394a3833393947acb8d504f686f136932499fbad8df5b0076586d2ccc"} Jan 29 15:45:01 crc kubenswrapper[4767]: I0129 15:45:01.717956 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" event={"ID":"1bac5b3d-af3b-489d-a798-10e2e5d6e221","Type":"ContainerStarted","Data":"e88288b0393c952c13a3c6687ba5ebdde311c1cea9be77bb1b794cbff94b2054"} Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.008609 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.087228 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bac5b3d-af3b-489d-a798-10e2e5d6e221-config-volume\") pod \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.087285 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5tpw\" (UniqueName: \"kubernetes.io/projected/1bac5b3d-af3b-489d-a798-10e2e5d6e221-kube-api-access-q5tpw\") pod \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.087386 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bac5b3d-af3b-489d-a798-10e2e5d6e221-secret-volume\") pod \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\" (UID: \"1bac5b3d-af3b-489d-a798-10e2e5d6e221\") " Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.088041 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bac5b3d-af3b-489d-a798-10e2e5d6e221-config-volume" (OuterVolumeSpecName: "config-volume") pod "1bac5b3d-af3b-489d-a798-10e2e5d6e221" (UID: "1bac5b3d-af3b-489d-a798-10e2e5d6e221"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.089025 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1bac5b3d-af3b-489d-a798-10e2e5d6e221-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.093135 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bac5b3d-af3b-489d-a798-10e2e5d6e221-kube-api-access-q5tpw" (OuterVolumeSpecName: "kube-api-access-q5tpw") pod "1bac5b3d-af3b-489d-a798-10e2e5d6e221" (UID: "1bac5b3d-af3b-489d-a798-10e2e5d6e221"). InnerVolumeSpecName "kube-api-access-q5tpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.094079 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bac5b3d-af3b-489d-a798-10e2e5d6e221-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1bac5b3d-af3b-489d-a798-10e2e5d6e221" (UID: "1bac5b3d-af3b-489d-a798-10e2e5d6e221"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.190913 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5tpw\" (UniqueName: \"kubernetes.io/projected/1bac5b3d-af3b-489d-a798-10e2e5d6e221-kube-api-access-q5tpw\") on node \"crc\" DevicePath \"\"" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.190968 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1bac5b3d-af3b-489d-a798-10e2e5d6e221-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.742295 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" event={"ID":"1bac5b3d-af3b-489d-a798-10e2e5d6e221","Type":"ContainerDied","Data":"e88288b0393c952c13a3c6687ba5ebdde311c1cea9be77bb1b794cbff94b2054"} Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.742342 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e88288b0393c952c13a3c6687ba5ebdde311c1cea9be77bb1b794cbff94b2054" Jan 29 15:45:03 crc kubenswrapper[4767]: I0129 15:45:03.742349 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495025-h4v8n" Jan 29 15:45:04 crc kubenswrapper[4767]: I0129 15:45:04.092395 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648"] Jan 29 15:45:04 crc kubenswrapper[4767]: I0129 15:45:04.099708 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494980-lr648"] Jan 29 15:45:05 crc kubenswrapper[4767]: I0129 15:45:05.027597 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c6ee408-b89d-4edf-b18f-d139046e8ccf" path="/var/lib/kubelet/pods/8c6ee408-b89d-4edf-b18f-d139046e8ccf/volumes" Jan 29 15:45:08 crc kubenswrapper[4767]: E0129 15:45:08.021965 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:45:11 crc kubenswrapper[4767]: I0129 15:45:11.019008 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:45:11 crc kubenswrapper[4767]: E0129 15:45:11.019935 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:45:21 crc kubenswrapper[4767]: E0129 15:45:21.021131 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:45:24 crc kubenswrapper[4767]: I0129 15:45:24.018866 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:45:24 crc kubenswrapper[4767]: E0129 15:45:24.019456 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:45:33 crc kubenswrapper[4767]: E0129 15:45:33.020397 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:45:38 crc kubenswrapper[4767]: I0129 15:45:38.017924 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:45:38 crc kubenswrapper[4767]: E0129 15:45:38.018636 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:45:46 crc kubenswrapper[4767]: I0129 15:45:46.097406 4767 scope.go:117] "RemoveContainer" containerID="91fc9c16f67dbc1338dd226b5538237a4ce9492154c22c51dd7a5241b9b254c7" Jan 29 15:45:47 crc kubenswrapper[4767]: E0129 15:45:47.020528 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:45:49 crc kubenswrapper[4767]: I0129 15:45:49.018512 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:45:50 crc kubenswrapper[4767]: I0129 15:45:50.184187 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"7bf68409fd71c87a32f36bd9173d32e10c1e02a78dde29c859e760534a1f6c9e"} Jan 29 15:46:01 crc kubenswrapper[4767]: E0129 15:46:01.022197 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:46:13 crc kubenswrapper[4767]: E0129 15:46:13.021069 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:46:26 crc kubenswrapper[4767]: E0129 15:46:26.022541 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:46:37 crc kubenswrapper[4767]: E0129 15:46:37.021573 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:46:50 crc kubenswrapper[4767]: E0129 15:46:50.020535 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:47:02 crc kubenswrapper[4767]: E0129 15:47:02.021306 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:47:15 crc kubenswrapper[4767]: E0129 15:47:15.026283 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:47:26 crc kubenswrapper[4767]: E0129 15:47:26.164211 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:47:26 crc kubenswrapper[4767]: E0129 15:47:26.165313 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75pk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsqfw_openshift-marketplace(405e1951-f911-428a-ad5f-2f1f29d76506): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:47:26 crc kubenswrapper[4767]: E0129 15:47:26.167970 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.358831 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hk8fs"] Jan 29 15:47:28 crc kubenswrapper[4767]: E0129 15:47:28.360071 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bac5b3d-af3b-489d-a798-10e2e5d6e221" containerName="collect-profiles" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.360088 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bac5b3d-af3b-489d-a798-10e2e5d6e221" containerName="collect-profiles" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.360226 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bac5b3d-af3b-489d-a798-10e2e5d6e221" containerName="collect-profiles" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.361217 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.370948 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hk8fs"] Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.412021 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zprd\" (UniqueName: \"kubernetes.io/projected/97a3c218-d649-406f-94e9-99a049ca6077-kube-api-access-4zprd\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.412182 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-catalog-content\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.412212 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-utilities\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.513973 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-catalog-content\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.514052 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-utilities\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.514107 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zprd\" (UniqueName: \"kubernetes.io/projected/97a3c218-d649-406f-94e9-99a049ca6077-kube-api-access-4zprd\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.514658 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-catalog-content\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.514794 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-utilities\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.547962 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zprd\" (UniqueName: \"kubernetes.io/projected/97a3c218-d649-406f-94e9-99a049ca6077-kube-api-access-4zprd\") pod \"certified-operators-hk8fs\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:28 crc kubenswrapper[4767]: I0129 15:47:28.695605 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:29 crc kubenswrapper[4767]: I0129 15:47:29.163747 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hk8fs"] Jan 29 15:47:29 crc kubenswrapper[4767]: W0129 15:47:29.171209 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97a3c218_d649_406f_94e9_99a049ca6077.slice/crio-2420d377d6f10332ed315bb072773797dfc8a6de0a095d31913237d9a9731436 WatchSource:0}: Error finding container 2420d377d6f10332ed315bb072773797dfc8a6de0a095d31913237d9a9731436: Status 404 returned error can't find the container with id 2420d377d6f10332ed315bb072773797dfc8a6de0a095d31913237d9a9731436 Jan 29 15:47:30 crc kubenswrapper[4767]: I0129 15:47:30.080035 4767 generic.go:334] "Generic (PLEG): container finished" podID="97a3c218-d649-406f-94e9-99a049ca6077" containerID="49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842" exitCode=0 Jan 29 15:47:30 crc kubenswrapper[4767]: I0129 15:47:30.080080 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk8fs" event={"ID":"97a3c218-d649-406f-94e9-99a049ca6077","Type":"ContainerDied","Data":"49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842"} Jan 29 15:47:30 crc kubenswrapper[4767]: I0129 15:47:30.080107 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk8fs" event={"ID":"97a3c218-d649-406f-94e9-99a049ca6077","Type":"ContainerStarted","Data":"2420d377d6f10332ed315bb072773797dfc8a6de0a095d31913237d9a9731436"} Jan 29 15:47:33 crc kubenswrapper[4767]: I0129 15:47:33.112947 4767 generic.go:334] "Generic (PLEG): container finished" podID="97a3c218-d649-406f-94e9-99a049ca6077" containerID="84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc" exitCode=0 Jan 29 15:47:33 crc kubenswrapper[4767]: I0129 15:47:33.113004 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk8fs" event={"ID":"97a3c218-d649-406f-94e9-99a049ca6077","Type":"ContainerDied","Data":"84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc"} Jan 29 15:47:35 crc kubenswrapper[4767]: I0129 15:47:35.133090 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk8fs" event={"ID":"97a3c218-d649-406f-94e9-99a049ca6077","Type":"ContainerStarted","Data":"08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d"} Jan 29 15:47:36 crc kubenswrapper[4767]: I0129 15:47:36.183611 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hk8fs" podStartSLOduration=3.7190719679999997 podStartE2EDuration="8.183590796s" podCreationTimestamp="2026-01-29 15:47:28 +0000 UTC" firstStartedPulling="2026-01-29 15:47:30.082167098 +0000 UTC m=+2805.892484137" lastFinishedPulling="2026-01-29 15:47:34.546685886 +0000 UTC m=+2810.357002965" observedRunningTime="2026-01-29 15:47:36.177178593 +0000 UTC m=+2811.987495632" watchObservedRunningTime="2026-01-29 15:47:36.183590796 +0000 UTC m=+2811.993907845" Jan 29 15:47:38 crc kubenswrapper[4767]: E0129 15:47:38.024217 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:47:38 crc kubenswrapper[4767]: I0129 15:47:38.696819 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:38 crc kubenswrapper[4767]: I0129 15:47:38.697202 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:38 crc kubenswrapper[4767]: I0129 15:47:38.746321 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:48 crc kubenswrapper[4767]: I0129 15:47:48.735747 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:50 crc kubenswrapper[4767]: E0129 15:47:50.022528 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:47:51 crc kubenswrapper[4767]: I0129 15:47:51.192522 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hk8fs"] Jan 29 15:47:51 crc kubenswrapper[4767]: I0129 15:47:51.192925 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hk8fs" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="registry-server" containerID="cri-o://08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d" gracePeriod=2 Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.152838 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.209195 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zprd\" (UniqueName: \"kubernetes.io/projected/97a3c218-d649-406f-94e9-99a049ca6077-kube-api-access-4zprd\") pod \"97a3c218-d649-406f-94e9-99a049ca6077\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.209283 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-utilities\") pod \"97a3c218-d649-406f-94e9-99a049ca6077\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.209303 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-catalog-content\") pod \"97a3c218-d649-406f-94e9-99a049ca6077\" (UID: \"97a3c218-d649-406f-94e9-99a049ca6077\") " Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.209986 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-utilities" (OuterVolumeSpecName: "utilities") pod "97a3c218-d649-406f-94e9-99a049ca6077" (UID: "97a3c218-d649-406f-94e9-99a049ca6077"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.215097 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a3c218-d649-406f-94e9-99a049ca6077-kube-api-access-4zprd" (OuterVolumeSpecName: "kube-api-access-4zprd") pod "97a3c218-d649-406f-94e9-99a049ca6077" (UID: "97a3c218-d649-406f-94e9-99a049ca6077"). InnerVolumeSpecName "kube-api-access-4zprd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.255963 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97a3c218-d649-406f-94e9-99a049ca6077" (UID: "97a3c218-d649-406f-94e9-99a049ca6077"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.284451 4767 generic.go:334] "Generic (PLEG): container finished" podID="97a3c218-d649-406f-94e9-99a049ca6077" containerID="08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d" exitCode=0 Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.284509 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk8fs" event={"ID":"97a3c218-d649-406f-94e9-99a049ca6077","Type":"ContainerDied","Data":"08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d"} Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.284512 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk8fs" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.284535 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk8fs" event={"ID":"97a3c218-d649-406f-94e9-99a049ca6077","Type":"ContainerDied","Data":"2420d377d6f10332ed315bb072773797dfc8a6de0a095d31913237d9a9731436"} Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.284552 4767 scope.go:117] "RemoveContainer" containerID="08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.304785 4767 scope.go:117] "RemoveContainer" containerID="84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.314492 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zprd\" (UniqueName: \"kubernetes.io/projected/97a3c218-d649-406f-94e9-99a049ca6077-kube-api-access-4zprd\") on node \"crc\" DevicePath \"\"" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.314543 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.314674 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a3c218-d649-406f-94e9-99a049ca6077-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.323977 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hk8fs"] Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.331811 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hk8fs"] Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.355661 4767 scope.go:117] "RemoveContainer" containerID="49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.379148 4767 scope.go:117] "RemoveContainer" containerID="08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d" Jan 29 15:47:52 crc kubenswrapper[4767]: E0129 15:47:52.381772 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d\": container with ID starting with 08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d not found: ID does not exist" containerID="08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.381814 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d"} err="failed to get container status \"08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d\": rpc error: code = NotFound desc = could not find container \"08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d\": container with ID starting with 08f8eb553b4f4f4f2e6c1f8f261c567db08d3ea128cdcb32e24019735ad7052d not found: ID does not exist" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.381841 4767 scope.go:117] "RemoveContainer" containerID="84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc" Jan 29 15:47:52 crc kubenswrapper[4767]: E0129 15:47:52.382519 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc\": container with ID starting with 84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc not found: ID does not exist" containerID="84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.382549 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc"} err="failed to get container status \"84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc\": rpc error: code = NotFound desc = could not find container \"84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc\": container with ID starting with 84013d397900b6760e3cd90c4a5c4ca2b59539f015efa73a24c98e0ee122f3bc not found: ID does not exist" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.382568 4767 scope.go:117] "RemoveContainer" containerID="49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842" Jan 29 15:47:52 crc kubenswrapper[4767]: E0129 15:47:52.383080 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842\": container with ID starting with 49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842 not found: ID does not exist" containerID="49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842" Jan 29 15:47:52 crc kubenswrapper[4767]: I0129 15:47:52.383111 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842"} err="failed to get container status \"49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842\": rpc error: code = NotFound desc = could not find container \"49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842\": container with ID starting with 49bbd488e5754eeff09907a760bf17b21c1846c5e46b008321fd2d39d5140842 not found: ID does not exist" Jan 29 15:47:53 crc kubenswrapper[4767]: I0129 15:47:53.038882 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a3c218-d649-406f-94e9-99a049ca6077" path="/var/lib/kubelet/pods/97a3c218-d649-406f-94e9-99a049ca6077/volumes" Jan 29 15:48:01 crc kubenswrapper[4767]: E0129 15:48:01.022012 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:48:12 crc kubenswrapper[4767]: I0129 15:48:12.806412 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:48:12 crc kubenswrapper[4767]: I0129 15:48:12.807074 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:48:14 crc kubenswrapper[4767]: E0129 15:48:14.020297 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:48:26 crc kubenswrapper[4767]: E0129 15:48:26.021335 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:48:41 crc kubenswrapper[4767]: E0129 15:48:41.021786 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:48:42 crc kubenswrapper[4767]: I0129 15:48:42.805879 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:48:42 crc kubenswrapper[4767]: I0129 15:48:42.805964 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:48:54 crc kubenswrapper[4767]: E0129 15:48:54.021353 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:49:06 crc kubenswrapper[4767]: E0129 15:49:06.022847 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:49:12 crc kubenswrapper[4767]: I0129 15:49:12.806235 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:49:12 crc kubenswrapper[4767]: I0129 15:49:12.806789 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:49:12 crc kubenswrapper[4767]: I0129 15:49:12.806852 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:49:12 crc kubenswrapper[4767]: I0129 15:49:12.807551 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7bf68409fd71c87a32f36bd9173d32e10c1e02a78dde29c859e760534a1f6c9e"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:49:12 crc kubenswrapper[4767]: I0129 15:49:12.807618 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://7bf68409fd71c87a32f36bd9173d32e10c1e02a78dde29c859e760534a1f6c9e" gracePeriod=600 Jan 29 15:49:13 crc kubenswrapper[4767]: I0129 15:49:13.883347 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="7bf68409fd71c87a32f36bd9173d32e10c1e02a78dde29c859e760534a1f6c9e" exitCode=0 Jan 29 15:49:13 crc kubenswrapper[4767]: I0129 15:49:13.883460 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"7bf68409fd71c87a32f36bd9173d32e10c1e02a78dde29c859e760534a1f6c9e"} Jan 29 15:49:13 crc kubenswrapper[4767]: I0129 15:49:13.883821 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e"} Jan 29 15:49:13 crc kubenswrapper[4767]: I0129 15:49:13.883853 4767 scope.go:117] "RemoveContainer" containerID="2f26bdf03b35897ce51c0c5549a09b5947b60252d668fdbfe9a527dd86374cf4" Jan 29 15:49:18 crc kubenswrapper[4767]: E0129 15:49:18.021401 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:49:32 crc kubenswrapper[4767]: E0129 15:49:32.022586 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:49:47 crc kubenswrapper[4767]: E0129 15:49:47.022039 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:50:00 crc kubenswrapper[4767]: E0129 15:50:00.023237 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:50:12 crc kubenswrapper[4767]: E0129 15:50:12.021586 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:50:24 crc kubenswrapper[4767]: E0129 15:50:24.020690 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:50:36 crc kubenswrapper[4767]: E0129 15:50:36.021067 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.001520 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2gjdb"] Jan 29 15:50:42 crc kubenswrapper[4767]: E0129 15:50:42.002038 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="extract-content" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.002052 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="extract-content" Jan 29 15:50:42 crc kubenswrapper[4767]: E0129 15:50:42.002067 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="registry-server" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.002076 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="registry-server" Jan 29 15:50:42 crc kubenswrapper[4767]: E0129 15:50:42.002085 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="extract-utilities" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.002092 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="extract-utilities" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.002763 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a3c218-d649-406f-94e9-99a049ca6077" containerName="registry-server" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.003729 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.028183 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gjdb"] Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.110684 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-utilities\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.110763 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxfgs\" (UniqueName: \"kubernetes.io/projected/39267376-c247-4c9a-b661-b860cb093da0-kube-api-access-hxfgs\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.110816 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-catalog-content\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.212211 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-utilities\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.212346 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxfgs\" (UniqueName: \"kubernetes.io/projected/39267376-c247-4c9a-b661-b860cb093da0-kube-api-access-hxfgs\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.212798 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-utilities\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.212879 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-catalog-content\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.213398 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-catalog-content\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.233948 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxfgs\" (UniqueName: \"kubernetes.io/projected/39267376-c247-4c9a-b661-b860cb093da0-kube-api-access-hxfgs\") pod \"redhat-marketplace-2gjdb\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.323535 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:50:42 crc kubenswrapper[4767]: I0129 15:50:42.838104 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gjdb"] Jan 29 15:50:43 crc kubenswrapper[4767]: I0129 15:50:43.635783 4767 generic.go:334] "Generic (PLEG): container finished" podID="39267376-c247-4c9a-b661-b860cb093da0" containerID="7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb" exitCode=0 Jan 29 15:50:43 crc kubenswrapper[4767]: I0129 15:50:43.635979 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gjdb" event={"ID":"39267376-c247-4c9a-b661-b860cb093da0","Type":"ContainerDied","Data":"7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb"} Jan 29 15:50:43 crc kubenswrapper[4767]: I0129 15:50:43.636224 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gjdb" event={"ID":"39267376-c247-4c9a-b661-b860cb093da0","Type":"ContainerStarted","Data":"1e395cd9ee8b8d25961e2824490a22adae708ce29cb63b7d83fcbcf5a810efb9"} Jan 29 15:50:43 crc kubenswrapper[4767]: I0129 15:50:43.640105 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:50:43 crc kubenswrapper[4767]: E0129 15:50:43.785222 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:50:43 crc kubenswrapper[4767]: E0129 15:50:43.785533 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hxfgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2gjdb_openshift-marketplace(39267376-c247-4c9a-b661-b860cb093da0): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:50:43 crc kubenswrapper[4767]: E0129 15:50:43.786748 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:50:44 crc kubenswrapper[4767]: E0129 15:50:44.645992 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:50:47 crc kubenswrapper[4767]: E0129 15:50:47.031985 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:50:59 crc kubenswrapper[4767]: E0129 15:50:59.020671 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:50:59 crc kubenswrapper[4767]: E0129 15:50:59.150449 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:50:59 crc kubenswrapper[4767]: E0129 15:50:59.151215 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hxfgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2gjdb_openshift-marketplace(39267376-c247-4c9a-b661-b860cb093da0): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:50:59 crc kubenswrapper[4767]: E0129 15:50:59.152551 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:51:10 crc kubenswrapper[4767]: E0129 15:51:10.021311 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:51:10 crc kubenswrapper[4767]: E0129 15:51:10.021312 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:51:21 crc kubenswrapper[4767]: E0129 15:51:21.151546 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:51:21 crc kubenswrapper[4767]: E0129 15:51:21.152909 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hxfgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2gjdb_openshift-marketplace(39267376-c247-4c9a-b661-b860cb093da0): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:51:21 crc kubenswrapper[4767]: E0129 15:51:21.154184 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:51:25 crc kubenswrapper[4767]: E0129 15:51:25.025676 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:51:36 crc kubenswrapper[4767]: E0129 15:51:36.021994 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:51:38 crc kubenswrapper[4767]: E0129 15:51:38.022580 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:51:42 crc kubenswrapper[4767]: I0129 15:51:42.805826 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:51:42 crc kubenswrapper[4767]: I0129 15:51:42.806311 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:51:51 crc kubenswrapper[4767]: E0129 15:51:51.020079 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:51:53 crc kubenswrapper[4767]: E0129 15:51:53.029318 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:52:03 crc kubenswrapper[4767]: E0129 15:52:03.165086 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 29 15:52:03 crc kubenswrapper[4767]: E0129 15:52:03.165684 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hxfgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2gjdb_openshift-marketplace(39267376-c247-4c9a-b661-b860cb093da0): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:52:03 crc kubenswrapper[4767]: E0129 15:52:03.167025 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-marketplace-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:52:04 crc kubenswrapper[4767]: E0129 15:52:04.021592 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:52:12 crc kubenswrapper[4767]: I0129 15:52:12.805204 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:52:12 crc kubenswrapper[4767]: I0129 15:52:12.805625 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:52:14 crc kubenswrapper[4767]: E0129 15:52:14.020405 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:52:17 crc kubenswrapper[4767]: E0129 15:52:17.020693 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:52:25 crc kubenswrapper[4767]: E0129 15:52:25.028635 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:52:28 crc kubenswrapper[4767]: E0129 15:52:28.149407 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:52:28 crc kubenswrapper[4767]: E0129 15:52:28.150329 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-75pk6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xsqfw_openshift-marketplace(405e1951-f911-428a-ad5f-2f1f29d76506): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:52:28 crc kubenswrapper[4767]: E0129 15:52:28.151666 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.254047 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4ccj7"] Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.255588 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.267847 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-catalog-content\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.267909 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-utilities\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.267955 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6ztz\" (UniqueName: \"kubernetes.io/projected/931bfdcc-a964-47e2-a23f-456dcc81061b-kube-api-access-k6ztz\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.270104 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ccj7"] Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.369649 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-catalog-content\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.369717 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-utilities\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.369763 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6ztz\" (UniqueName: \"kubernetes.io/projected/931bfdcc-a964-47e2-a23f-456dcc81061b-kube-api-access-k6ztz\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.370260 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-catalog-content\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.370712 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-utilities\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.393670 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6ztz\" (UniqueName: \"kubernetes.io/projected/931bfdcc-a964-47e2-a23f-456dcc81061b-kube-api-access-k6ztz\") pod \"redhat-operators-4ccj7\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:28 crc kubenswrapper[4767]: I0129 15:52:28.579934 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:52:29 crc kubenswrapper[4767]: I0129 15:52:29.027815 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ccj7"] Jan 29 15:52:29 crc kubenswrapper[4767]: I0129 15:52:29.452362 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerStarted","Data":"ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137"} Jan 29 15:52:29 crc kubenswrapper[4767]: I0129 15:52:29.452670 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerStarted","Data":"33a8e6ef2e4ccecff4007df36c49136411da8259f5f46ac5f5b05f650e092b10"} Jan 29 15:52:30 crc kubenswrapper[4767]: I0129 15:52:30.484122 4767 generic.go:334] "Generic (PLEG): container finished" podID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerID="ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137" exitCode=0 Jan 29 15:52:30 crc kubenswrapper[4767]: I0129 15:52:30.485002 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerDied","Data":"ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137"} Jan 29 15:52:30 crc kubenswrapper[4767]: E0129 15:52:30.606636 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:52:30 crc kubenswrapper[4767]: E0129 15:52:30.606781 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k6ztz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4ccj7_openshift-marketplace(931bfdcc-a964-47e2-a23f-456dcc81061b): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:52:30 crc kubenswrapper[4767]: E0129 15:52:30.607972 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-4ccj7" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" Jan 29 15:52:31 crc kubenswrapper[4767]: E0129 15:52:31.491716 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4ccj7" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" Jan 29 15:52:39 crc kubenswrapper[4767]: E0129 15:52:39.020621 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:52:42 crc kubenswrapper[4767]: I0129 15:52:42.806580 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 15:52:42 crc kubenswrapper[4767]: I0129 15:52:42.807015 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 15:52:42 crc kubenswrapper[4767]: I0129 15:52:42.807074 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 15:52:42 crc kubenswrapper[4767]: I0129 15:52:42.808101 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 15:52:42 crc kubenswrapper[4767]: I0129 15:52:42.808204 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" gracePeriod=600 Jan 29 15:52:43 crc kubenswrapper[4767]: E0129 15:52:43.019640 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:52:43 crc kubenswrapper[4767]: E0129 15:52:43.147692 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 15:52:43 crc kubenswrapper[4767]: E0129 15:52:43.147850 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k6ztz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4ccj7_openshift-marketplace(931bfdcc-a964-47e2-a23f-456dcc81061b): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:52:43 crc kubenswrapper[4767]: E0129 15:52:43.149127 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-4ccj7" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" Jan 29 15:52:43 crc kubenswrapper[4767]: E0129 15:52:43.474529 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:52:43 crc kubenswrapper[4767]: I0129 15:52:43.568851 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" exitCode=0 Jan 29 15:52:43 crc kubenswrapper[4767]: I0129 15:52:43.568933 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e"} Jan 29 15:52:43 crc kubenswrapper[4767]: I0129 15:52:43.569361 4767 scope.go:117] "RemoveContainer" containerID="7bf68409fd71c87a32f36bd9173d32e10c1e02a78dde29c859e760534a1f6c9e" Jan 29 15:52:43 crc kubenswrapper[4767]: I0129 15:52:43.569871 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:52:43 crc kubenswrapper[4767]: E0129 15:52:43.570100 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:52:53 crc kubenswrapper[4767]: E0129 15:52:53.021494 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:52:54 crc kubenswrapper[4767]: I0129 15:52:54.018120 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:52:54 crc kubenswrapper[4767]: E0129 15:52:54.018777 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:52:55 crc kubenswrapper[4767]: E0129 15:52:55.035746 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4ccj7" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" Jan 29 15:52:58 crc kubenswrapper[4767]: E0129 15:52:58.021964 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:53:07 crc kubenswrapper[4767]: E0129 15:53:07.020724 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:53:09 crc kubenswrapper[4767]: I0129 15:53:09.019184 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:53:09 crc kubenswrapper[4767]: E0129 15:53:09.019436 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:53:11 crc kubenswrapper[4767]: I0129 15:53:11.777309 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerStarted","Data":"706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290"} Jan 29 15:53:12 crc kubenswrapper[4767]: I0129 15:53:12.787694 4767 generic.go:334] "Generic (PLEG): container finished" podID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerID="706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290" exitCode=0 Jan 29 15:53:12 crc kubenswrapper[4767]: I0129 15:53:12.787744 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerDied","Data":"706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290"} Jan 29 15:53:13 crc kubenswrapper[4767]: E0129 15:53:13.019322 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:53:13 crc kubenswrapper[4767]: I0129 15:53:13.797244 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerStarted","Data":"d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004"} Jan 29 15:53:13 crc kubenswrapper[4767]: I0129 15:53:13.817310 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4ccj7" podStartSLOduration=3.063893126 podStartE2EDuration="45.817289877s" podCreationTimestamp="2026-01-29 15:52:28 +0000 UTC" firstStartedPulling="2026-01-29 15:52:30.485656505 +0000 UTC m=+3106.295973544" lastFinishedPulling="2026-01-29 15:53:13.239053256 +0000 UTC m=+3149.049370295" observedRunningTime="2026-01-29 15:53:13.816051331 +0000 UTC m=+3149.626368370" watchObservedRunningTime="2026-01-29 15:53:13.817289877 +0000 UTC m=+3149.627606916" Jan 29 15:53:18 crc kubenswrapper[4767]: I0129 15:53:18.593652 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:53:18 crc kubenswrapper[4767]: I0129 15:53:18.594141 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:53:19 crc kubenswrapper[4767]: I0129 15:53:19.648653 4767 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4ccj7" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="registry-server" probeResult="failure" output=< Jan 29 15:53:19 crc kubenswrapper[4767]: timeout: failed to connect service ":50051" within 1s Jan 29 15:53:19 crc kubenswrapper[4767]: > Jan 29 15:53:20 crc kubenswrapper[4767]: I0129 15:53:20.018366 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:53:20 crc kubenswrapper[4767]: E0129 15:53:20.018744 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:53:21 crc kubenswrapper[4767]: E0129 15:53:21.020884 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" Jan 29 15:53:26 crc kubenswrapper[4767]: E0129 15:53:26.021034 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:53:28 crc kubenswrapper[4767]: I0129 15:53:28.656946 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:53:28 crc kubenswrapper[4767]: I0129 15:53:28.720504 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:53:29 crc kubenswrapper[4767]: I0129 15:53:29.501461 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ccj7"] Jan 29 15:53:29 crc kubenswrapper[4767]: I0129 15:53:29.917766 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4ccj7" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="registry-server" containerID="cri-o://d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004" gracePeriod=2 Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.890162 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.926330 4767 generic.go:334] "Generic (PLEG): container finished" podID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerID="d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004" exitCode=0 Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.926383 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerDied","Data":"d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004"} Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.926419 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ccj7" event={"ID":"931bfdcc-a964-47e2-a23f-456dcc81061b","Type":"ContainerDied","Data":"33a8e6ef2e4ccecff4007df36c49136411da8259f5f46ac5f5b05f650e092b10"} Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.926480 4767 scope.go:117] "RemoveContainer" containerID="d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.926681 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ccj7" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.946381 4767 scope.go:117] "RemoveContainer" containerID="706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.964190 4767 scope.go:117] "RemoveContainer" containerID="ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.987357 4767 scope.go:117] "RemoveContainer" containerID="d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004" Jan 29 15:53:30 crc kubenswrapper[4767]: E0129 15:53:30.987753 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004\": container with ID starting with d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004 not found: ID does not exist" containerID="d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.987792 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004"} err="failed to get container status \"d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004\": rpc error: code = NotFound desc = could not find container \"d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004\": container with ID starting with d8af2b37b8a5205fe391483dfd71bf9f828b6bb40433c02a7a0ea26f63ed7004 not found: ID does not exist" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.987833 4767 scope.go:117] "RemoveContainer" containerID="706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290" Jan 29 15:53:30 crc kubenswrapper[4767]: E0129 15:53:30.988432 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290\": container with ID starting with 706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290 not found: ID does not exist" containerID="706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.988513 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290"} err="failed to get container status \"706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290\": rpc error: code = NotFound desc = could not find container \"706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290\": container with ID starting with 706af74941443cb136651040ca3b076be3d23a387c1dd5846a6577034f6c6290 not found: ID does not exist" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.988590 4767 scope.go:117] "RemoveContainer" containerID="ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137" Jan 29 15:53:30 crc kubenswrapper[4767]: E0129 15:53:30.989157 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137\": container with ID starting with ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137 not found: ID does not exist" containerID="ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137" Jan 29 15:53:30 crc kubenswrapper[4767]: I0129 15:53:30.989218 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137"} err="failed to get container status \"ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137\": rpc error: code = NotFound desc = could not find container \"ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137\": container with ID starting with ee44841337d24232850c3f170527fa6581dc216f71e1a7f055fe0899ef01f137 not found: ID does not exist" Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.014828 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6ztz\" (UniqueName: \"kubernetes.io/projected/931bfdcc-a964-47e2-a23f-456dcc81061b-kube-api-access-k6ztz\") pod \"931bfdcc-a964-47e2-a23f-456dcc81061b\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.014880 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-catalog-content\") pod \"931bfdcc-a964-47e2-a23f-456dcc81061b\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.014955 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-utilities\") pod \"931bfdcc-a964-47e2-a23f-456dcc81061b\" (UID: \"931bfdcc-a964-47e2-a23f-456dcc81061b\") " Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.016024 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-utilities" (OuterVolumeSpecName: "utilities") pod "931bfdcc-a964-47e2-a23f-456dcc81061b" (UID: "931bfdcc-a964-47e2-a23f-456dcc81061b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.020986 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/931bfdcc-a964-47e2-a23f-456dcc81061b-kube-api-access-k6ztz" (OuterVolumeSpecName: "kube-api-access-k6ztz") pod "931bfdcc-a964-47e2-a23f-456dcc81061b" (UID: "931bfdcc-a964-47e2-a23f-456dcc81061b"). InnerVolumeSpecName "kube-api-access-k6ztz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.117743 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6ztz\" (UniqueName: \"kubernetes.io/projected/931bfdcc-a964-47e2-a23f-456dcc81061b-kube-api-access-k6ztz\") on node \"crc\" DevicePath \"\"" Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.118171 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.128687 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "931bfdcc-a964-47e2-a23f-456dcc81061b" (UID: "931bfdcc-a964-47e2-a23f-456dcc81061b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.219821 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/931bfdcc-a964-47e2-a23f-456dcc81061b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.265941 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ccj7"] Jan 29 15:53:31 crc kubenswrapper[4767]: I0129 15:53:31.271110 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4ccj7"] Jan 29 15:53:32 crc kubenswrapper[4767]: I0129 15:53:32.018387 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:53:32 crc kubenswrapper[4767]: E0129 15:53:32.018784 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:53:33 crc kubenswrapper[4767]: I0129 15:53:33.028608 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" path="/var/lib/kubelet/pods/931bfdcc-a964-47e2-a23f-456dcc81061b/volumes" Jan 29 15:53:36 crc kubenswrapper[4767]: I0129 15:53:36.967607 4767 generic.go:334] "Generic (PLEG): container finished" podID="39267376-c247-4c9a-b661-b860cb093da0" containerID="ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca" exitCode=0 Jan 29 15:53:36 crc kubenswrapper[4767]: I0129 15:53:36.967670 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gjdb" event={"ID":"39267376-c247-4c9a-b661-b860cb093da0","Type":"ContainerDied","Data":"ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca"} Jan 29 15:53:37 crc kubenswrapper[4767]: I0129 15:53:37.977960 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gjdb" event={"ID":"39267376-c247-4c9a-b661-b860cb093da0","Type":"ContainerStarted","Data":"05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b"} Jan 29 15:53:38 crc kubenswrapper[4767]: I0129 15:53:38.005586 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2gjdb" podStartSLOduration=3.112921439 podStartE2EDuration="2m57.005553925s" podCreationTimestamp="2026-01-29 15:50:41 +0000 UTC" firstStartedPulling="2026-01-29 15:50:43.639672341 +0000 UTC m=+2999.449989400" lastFinishedPulling="2026-01-29 15:53:37.532304847 +0000 UTC m=+3173.342621886" observedRunningTime="2026-01-29 15:53:37.998424342 +0000 UTC m=+3173.808741381" watchObservedRunningTime="2026-01-29 15:53:38.005553925 +0000 UTC m=+3173.815870964" Jan 29 15:53:41 crc kubenswrapper[4767]: E0129 15:53:41.020808 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:53:42 crc kubenswrapper[4767]: I0129 15:53:42.324286 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:53:42 crc kubenswrapper[4767]: I0129 15:53:42.324428 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:53:42 crc kubenswrapper[4767]: I0129 15:53:42.370042 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:53:43 crc kubenswrapper[4767]: I0129 15:53:43.018755 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:53:43 crc kubenswrapper[4767]: E0129 15:53:43.019064 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:53:43 crc kubenswrapper[4767]: I0129 15:53:43.044828 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:53:43 crc kubenswrapper[4767]: I0129 15:53:43.288606 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gjdb"] Jan 29 15:53:45 crc kubenswrapper[4767]: I0129 15:53:45.027070 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2gjdb" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="registry-server" containerID="cri-o://05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b" gracePeriod=2 Jan 29 15:53:45 crc kubenswrapper[4767]: I0129 15:53:45.886026 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.034867 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-catalog-content\") pod \"39267376-c247-4c9a-b661-b860cb093da0\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.034943 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxfgs\" (UniqueName: \"kubernetes.io/projected/39267376-c247-4c9a-b661-b860cb093da0-kube-api-access-hxfgs\") pod \"39267376-c247-4c9a-b661-b860cb093da0\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.035011 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-utilities\") pod \"39267376-c247-4c9a-b661-b860cb093da0\" (UID: \"39267376-c247-4c9a-b661-b860cb093da0\") " Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.036140 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-utilities" (OuterVolumeSpecName: "utilities") pod "39267376-c247-4c9a-b661-b860cb093da0" (UID: "39267376-c247-4c9a-b661-b860cb093da0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.036955 4767 generic.go:334] "Generic (PLEG): container finished" podID="39267376-c247-4c9a-b661-b860cb093da0" containerID="05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b" exitCode=0 Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.036998 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gjdb" event={"ID":"39267376-c247-4c9a-b661-b860cb093da0","Type":"ContainerDied","Data":"05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b"} Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.037029 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2gjdb" event={"ID":"39267376-c247-4c9a-b661-b860cb093da0","Type":"ContainerDied","Data":"1e395cd9ee8b8d25961e2824490a22adae708ce29cb63b7d83fcbcf5a810efb9"} Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.037074 4767 scope.go:117] "RemoveContainer" containerID="05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.037101 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2gjdb" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.046037 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39267376-c247-4c9a-b661-b860cb093da0-kube-api-access-hxfgs" (OuterVolumeSpecName: "kube-api-access-hxfgs") pod "39267376-c247-4c9a-b661-b860cb093da0" (UID: "39267376-c247-4c9a-b661-b860cb093da0"). InnerVolumeSpecName "kube-api-access-hxfgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.069625 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39267376-c247-4c9a-b661-b860cb093da0" (UID: "39267376-c247-4c9a-b661-b860cb093da0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.093178 4767 scope.go:117] "RemoveContainer" containerID="ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.115033 4767 scope.go:117] "RemoveContainer" containerID="7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.131356 4767 scope.go:117] "RemoveContainer" containerID="05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b" Jan 29 15:53:46 crc kubenswrapper[4767]: E0129 15:53:46.132136 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b\": container with ID starting with 05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b not found: ID does not exist" containerID="05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.132273 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b"} err="failed to get container status \"05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b\": rpc error: code = NotFound desc = could not find container \"05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b\": container with ID starting with 05f48861bce2551ebb8b522f825189cceef781004d210ca372715385dc540e3b not found: ID does not exist" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.132373 4767 scope.go:117] "RemoveContainer" containerID="ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca" Jan 29 15:53:46 crc kubenswrapper[4767]: E0129 15:53:46.132914 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca\": container with ID starting with ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca not found: ID does not exist" containerID="ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.132998 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca"} err="failed to get container status \"ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca\": rpc error: code = NotFound desc = could not find container \"ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca\": container with ID starting with ee69ec88e8f362991660980709c308fcffd2823a2f0b4dec3a1728ee29f643ca not found: ID does not exist" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.133063 4767 scope.go:117] "RemoveContainer" containerID="7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb" Jan 29 15:53:46 crc kubenswrapper[4767]: E0129 15:53:46.133394 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb\": container with ID starting with 7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb not found: ID does not exist" containerID="7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.133476 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb"} err="failed to get container status \"7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb\": rpc error: code = NotFound desc = could not find container \"7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb\": container with ID starting with 7afb10d3e31ee0306613d25187eb5cd8408b2dacd1abaedee41bce4ec97593fb not found: ID does not exist" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.136272 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.136296 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxfgs\" (UniqueName: \"kubernetes.io/projected/39267376-c247-4c9a-b661-b860cb093da0-kube-api-access-hxfgs\") on node \"crc\" DevicePath \"\"" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.136307 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39267376-c247-4c9a-b661-b860cb093da0-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.365959 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gjdb"] Jan 29 15:53:46 crc kubenswrapper[4767]: I0129 15:53:46.373662 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2gjdb"] Jan 29 15:53:47 crc kubenswrapper[4767]: I0129 15:53:47.029887 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39267376-c247-4c9a-b661-b860cb093da0" path="/var/lib/kubelet/pods/39267376-c247-4c9a-b661-b860cb093da0/volumes" Jan 29 15:53:56 crc kubenswrapper[4767]: E0129 15:53:56.020978 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:53:57 crc kubenswrapper[4767]: I0129 15:53:57.019130 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:53:57 crc kubenswrapper[4767]: E0129 15:53:57.019783 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:54:09 crc kubenswrapper[4767]: I0129 15:54:09.018572 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:54:09 crc kubenswrapper[4767]: E0129 15:54:09.020451 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:54:11 crc kubenswrapper[4767]: E0129 15:54:11.038274 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:54:22 crc kubenswrapper[4767]: I0129 15:54:22.019010 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:54:22 crc kubenswrapper[4767]: E0129 15:54:22.019693 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:54:26 crc kubenswrapper[4767]: E0129 15:54:26.021510 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:54:33 crc kubenswrapper[4767]: I0129 15:54:33.018347 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:54:33 crc kubenswrapper[4767]: E0129 15:54:33.018842 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:54:39 crc kubenswrapper[4767]: E0129 15:54:39.021837 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:54:46 crc kubenswrapper[4767]: I0129 15:54:46.019394 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:54:46 crc kubenswrapper[4767]: E0129 15:54:46.020726 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:54:54 crc kubenswrapper[4767]: E0129 15:54:54.021610 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:54:57 crc kubenswrapper[4767]: I0129 15:54:57.019479 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:54:57 crc kubenswrapper[4767]: E0129 15:54:57.020198 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:55:09 crc kubenswrapper[4767]: E0129 15:55:09.021090 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:55:10 crc kubenswrapper[4767]: I0129 15:55:10.018624 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:55:10 crc kubenswrapper[4767]: E0129 15:55:10.019749 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:55:23 crc kubenswrapper[4767]: I0129 15:55:23.018416 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:55:23 crc kubenswrapper[4767]: E0129 15:55:23.019256 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:55:24 crc kubenswrapper[4767]: E0129 15:55:24.020630 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:55:36 crc kubenswrapper[4767]: I0129 15:55:36.018576 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:55:36 crc kubenswrapper[4767]: E0129 15:55:36.019313 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:55:37 crc kubenswrapper[4767]: E0129 15:55:37.020559 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:55:48 crc kubenswrapper[4767]: E0129 15:55:48.021002 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:55:49 crc kubenswrapper[4767]: I0129 15:55:49.019050 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:55:49 crc kubenswrapper[4767]: E0129 15:55:49.019297 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:55:59 crc kubenswrapper[4767]: E0129 15:55:59.021161 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:56:03 crc kubenswrapper[4767]: I0129 15:56:03.018189 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:56:03 crc kubenswrapper[4767]: E0129 15:56:03.018492 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:56:10 crc kubenswrapper[4767]: E0129 15:56:10.021666 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:56:17 crc kubenswrapper[4767]: I0129 15:56:17.019406 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:56:17 crc kubenswrapper[4767]: E0129 15:56:17.020304 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:56:25 crc kubenswrapper[4767]: E0129 15:56:25.030024 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:56:31 crc kubenswrapper[4767]: I0129 15:56:31.019205 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:56:31 crc kubenswrapper[4767]: E0129 15:56:31.020257 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:56:40 crc kubenswrapper[4767]: E0129 15:56:40.021527 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:56:46 crc kubenswrapper[4767]: I0129 15:56:46.018875 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:56:46 crc kubenswrapper[4767]: E0129 15:56:46.020622 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:56:52 crc kubenswrapper[4767]: E0129 15:56:52.025730 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:56:58 crc kubenswrapper[4767]: I0129 15:56:58.018846 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:56:58 crc kubenswrapper[4767]: E0129 15:56:58.019531 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:57:07 crc kubenswrapper[4767]: E0129 15:57:07.020018 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:57:13 crc kubenswrapper[4767]: I0129 15:57:13.019209 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:57:13 crc kubenswrapper[4767]: E0129 15:57:13.019672 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:57:22 crc kubenswrapper[4767]: E0129 15:57:22.021940 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" Jan 29 15:57:26 crc kubenswrapper[4767]: I0129 15:57:26.018994 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:57:26 crc kubenswrapper[4767]: E0129 15:57:26.019608 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:57:33 crc kubenswrapper[4767]: I0129 15:57:33.022026 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 15:57:35 crc kubenswrapper[4767]: I0129 15:57:35.187335 4767 generic.go:334] "Generic (PLEG): container finished" podID="405e1951-f911-428a-ad5f-2f1f29d76506" containerID="501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da" exitCode=0 Jan 29 15:57:35 crc kubenswrapper[4767]: I0129 15:57:35.187427 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsqfw" event={"ID":"405e1951-f911-428a-ad5f-2f1f29d76506","Type":"ContainerDied","Data":"501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da"} Jan 29 15:57:36 crc kubenswrapper[4767]: I0129 15:57:36.204082 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsqfw" event={"ID":"405e1951-f911-428a-ad5f-2f1f29d76506","Type":"ContainerStarted","Data":"63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c"} Jan 29 15:57:36 crc kubenswrapper[4767]: I0129 15:57:36.230309 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xsqfw" podStartSLOduration=2.38646857 podStartE2EDuration="15m49.23029302s" podCreationTimestamp="2026-01-29 15:41:47 +0000 UTC" firstStartedPulling="2026-01-29 15:41:48.809443854 +0000 UTC m=+2464.619760893" lastFinishedPulling="2026-01-29 15:57:35.653268304 +0000 UTC m=+3411.463585343" observedRunningTime="2026-01-29 15:57:36.223376303 +0000 UTC m=+3412.033693352" watchObservedRunningTime="2026-01-29 15:57:36.23029302 +0000 UTC m=+3412.040610059" Jan 29 15:57:37 crc kubenswrapper[4767]: I0129 15:57:37.941980 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:57:37 crc kubenswrapper[4767]: I0129 15:57:37.942332 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:57:37 crc kubenswrapper[4767]: I0129 15:57:37.988940 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:57:38 crc kubenswrapper[4767]: I0129 15:57:38.019043 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:57:38 crc kubenswrapper[4767]: E0129 15:57:38.019407 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 15:57:47 crc kubenswrapper[4767]: I0129 15:57:47.980435 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:57:48 crc kubenswrapper[4767]: I0129 15:57:48.032114 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xsqfw"] Jan 29 15:57:48 crc kubenswrapper[4767]: I0129 15:57:48.299622 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xsqfw" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="registry-server" containerID="cri-o://63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c" gracePeriod=2 Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.303511 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.308865 4767 generic.go:334] "Generic (PLEG): container finished" podID="405e1951-f911-428a-ad5f-2f1f29d76506" containerID="63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c" exitCode=0 Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.308927 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsqfw" event={"ID":"405e1951-f911-428a-ad5f-2f1f29d76506","Type":"ContainerDied","Data":"63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c"} Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.308951 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xsqfw" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.308968 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xsqfw" event={"ID":"405e1951-f911-428a-ad5f-2f1f29d76506","Type":"ContainerDied","Data":"ce00a99434da761e3ea1459914058e3a88285d1f730d61f548f5d7074a3d0404"} Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.308997 4767 scope.go:117] "RemoveContainer" containerID="63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.331034 4767 scope.go:117] "RemoveContainer" containerID="501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.361285 4767 scope.go:117] "RemoveContainer" containerID="93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.375325 4767 scope.go:117] "RemoveContainer" containerID="63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c" Jan 29 15:57:49 crc kubenswrapper[4767]: E0129 15:57:49.375976 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c\": container with ID starting with 63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c not found: ID does not exist" containerID="63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.376046 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c"} err="failed to get container status \"63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c\": rpc error: code = NotFound desc = could not find container \"63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c\": container with ID starting with 63907dbdf1a620a56dcb14cb196458abfcd7fa47331b7b9583a1c2770fd1843c not found: ID does not exist" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.376074 4767 scope.go:117] "RemoveContainer" containerID="501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da" Jan 29 15:57:49 crc kubenswrapper[4767]: E0129 15:57:49.376331 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da\": container with ID starting with 501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da not found: ID does not exist" containerID="501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.376356 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da"} err="failed to get container status \"501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da\": rpc error: code = NotFound desc = could not find container \"501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da\": container with ID starting with 501cb2889ad8564f0832fb16d57efd570d708311c2b68fa6889310739069f3da not found: ID does not exist" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.376375 4767 scope.go:117] "RemoveContainer" containerID="93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75" Jan 29 15:57:49 crc kubenswrapper[4767]: E0129 15:57:49.376718 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75\": container with ID starting with 93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75 not found: ID does not exist" containerID="93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.376743 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75"} err="failed to get container status \"93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75\": rpc error: code = NotFound desc = could not find container \"93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75\": container with ID starting with 93f4f0e41dd56ca30566c15ed33cd54a1f165aff685ae7c0af98af3c8a2abd75 not found: ID does not exist" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.453102 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75pk6\" (UniqueName: \"kubernetes.io/projected/405e1951-f911-428a-ad5f-2f1f29d76506-kube-api-access-75pk6\") pod \"405e1951-f911-428a-ad5f-2f1f29d76506\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.453170 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-utilities\") pod \"405e1951-f911-428a-ad5f-2f1f29d76506\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.453206 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-catalog-content\") pod \"405e1951-f911-428a-ad5f-2f1f29d76506\" (UID: \"405e1951-f911-428a-ad5f-2f1f29d76506\") " Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.458880 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-utilities" (OuterVolumeSpecName: "utilities") pod "405e1951-f911-428a-ad5f-2f1f29d76506" (UID: "405e1951-f911-428a-ad5f-2f1f29d76506"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.464250 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/405e1951-f911-428a-ad5f-2f1f29d76506-kube-api-access-75pk6" (OuterVolumeSpecName: "kube-api-access-75pk6") pod "405e1951-f911-428a-ad5f-2f1f29d76506" (UID: "405e1951-f911-428a-ad5f-2f1f29d76506"). InnerVolumeSpecName "kube-api-access-75pk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.501033 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "405e1951-f911-428a-ad5f-2f1f29d76506" (UID: "405e1951-f911-428a-ad5f-2f1f29d76506"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.555079 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75pk6\" (UniqueName: \"kubernetes.io/projected/405e1951-f911-428a-ad5f-2f1f29d76506-kube-api-access-75pk6\") on node \"crc\" DevicePath \"\"" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.555113 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.555199 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/405e1951-f911-428a-ad5f-2f1f29d76506-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.650248 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xsqfw"] Jan 29 15:57:49 crc kubenswrapper[4767]: I0129 15:57:49.658063 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xsqfw"] Jan 29 15:57:50 crc kubenswrapper[4767]: I0129 15:57:50.018993 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.027584 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" path="/var/lib/kubelet/pods/405e1951-f911-428a-ad5f-2f1f29d76506/volumes" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.512941 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"55581c47200bca1e0d0702088fd9033cee9ab992eba215fd248062d21ef80f24"} Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515509 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w2zss"] Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515802 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515819 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515832 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="extract-utilities" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515838 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="extract-utilities" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515848 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="extract-content" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515855 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="extract-content" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515869 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515874 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515882 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="extract-content" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515887 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="extract-content" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515911 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="extract-content" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515917 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="extract-content" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515929 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="extract-utilities" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515935 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="extract-utilities" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515943 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515949 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: E0129 15:57:51.515957 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="extract-utilities" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.515963 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="extract-utilities" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.516083 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="405e1951-f911-428a-ad5f-2f1f29d76506" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.516095 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="39267376-c247-4c9a-b661-b860cb093da0" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.516103 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="931bfdcc-a964-47e2-a23f-456dcc81061b" containerName="registry-server" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.517039 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.531116 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2zss"] Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.683816 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqnnp\" (UniqueName: \"kubernetes.io/projected/5f7003ac-5dc0-4319-92e8-203585c220c9-kube-api-access-bqnnp\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.683877 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-catalog-content\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.683964 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-utilities\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.785007 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-utilities\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.785117 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqnnp\" (UniqueName: \"kubernetes.io/projected/5f7003ac-5dc0-4319-92e8-203585c220c9-kube-api-access-bqnnp\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.785154 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-catalog-content\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.785468 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-utilities\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.785529 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-catalog-content\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.810376 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqnnp\" (UniqueName: \"kubernetes.io/projected/5f7003ac-5dc0-4319-92e8-203585c220c9-kube-api-access-bqnnp\") pod \"community-operators-w2zss\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:51 crc kubenswrapper[4767]: I0129 15:57:51.842100 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zss" Jan 29 15:57:52 crc kubenswrapper[4767]: I0129 15:57:52.098120 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w2zss"] Jan 29 15:57:52 crc kubenswrapper[4767]: I0129 15:57:52.521312 4767 generic.go:334] "Generic (PLEG): container finished" podID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerID="6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c" exitCode=0 Jan 29 15:57:52 crc kubenswrapper[4767]: I0129 15:57:52.521359 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zss" event={"ID":"5f7003ac-5dc0-4319-92e8-203585c220c9","Type":"ContainerDied","Data":"6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c"} Jan 29 15:57:52 crc kubenswrapper[4767]: I0129 15:57:52.521388 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zss" event={"ID":"5f7003ac-5dc0-4319-92e8-203585c220c9","Type":"ContainerStarted","Data":"a3af4104f25b5f9acc1941219dcd69af57a64bf15445a696b7b70db330805c12"} Jan 29 15:57:52 crc kubenswrapper[4767]: E0129 15:57:52.696478 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:57:52 crc kubenswrapper[4767]: E0129 15:57:52.696675 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-w2zss_openshift-marketplace(5f7003ac-5dc0-4319-92e8-203585c220c9): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:57:52 crc kubenswrapper[4767]: E0129 15:57:52.698362 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.431602 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5zmq2"] Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.434105 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.441415 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5zmq2"] Jan 29 15:57:53 crc kubenswrapper[4767]: E0129 15:57:53.540354 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.609730 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-utilities\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.609989 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-catalog-content\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.610110 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvrz8\" (UniqueName: \"kubernetes.io/projected/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-kube-api-access-pvrz8\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.711627 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-catalog-content\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.712006 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvrz8\" (UniqueName: \"kubernetes.io/projected/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-kube-api-access-pvrz8\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.712052 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-utilities\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.712478 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-utilities\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.712453 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-catalog-content\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.731163 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvrz8\" (UniqueName: \"kubernetes.io/projected/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-kube-api-access-pvrz8\") pod \"certified-operators-5zmq2\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:53 crc kubenswrapper[4767]: I0129 15:57:53.753867 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 15:57:54 crc kubenswrapper[4767]: I0129 15:57:54.185489 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5zmq2"] Jan 29 15:57:54 crc kubenswrapper[4767]: I0129 15:57:54.543317 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5zmq2" event={"ID":"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b","Type":"ContainerStarted","Data":"50a2c587606b0ef14bd0b500be2a0a06e62acb3501f4bf04020c241f42cbb57e"} Jan 29 15:57:55 crc kubenswrapper[4767]: I0129 15:57:55.556063 4767 generic.go:334] "Generic (PLEG): container finished" podID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerID="c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3" exitCode=0 Jan 29 15:57:55 crc kubenswrapper[4767]: I0129 15:57:55.556198 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5zmq2" event={"ID":"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b","Type":"ContainerDied","Data":"c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3"} Jan 29 15:57:55 crc kubenswrapper[4767]: E0129 15:57:55.683334 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:57:55 crc kubenswrapper[4767]: E0129 15:57:55.683520 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pvrz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5zmq2_openshift-marketplace(8b375cb3-2015-4d3f-b0d0-1e883fca2e5b): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:57:55 crc kubenswrapper[4767]: E0129 15:57:55.684741 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:57:56 crc kubenswrapper[4767]: E0129 15:57:56.565231 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:58:08 crc kubenswrapper[4767]: E0129 15:58:08.151780 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:58:08 crc kubenswrapper[4767]: E0129 15:58:08.152425 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pvrz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5zmq2_openshift-marketplace(8b375cb3-2015-4d3f-b0d0-1e883fca2e5b): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:58:08 crc kubenswrapper[4767]: E0129 15:58:08.153527 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:58:09 crc kubenswrapper[4767]: E0129 15:58:09.155826 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:58:09 crc kubenswrapper[4767]: E0129 15:58:09.156319 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-w2zss_openshift-marketplace(5f7003ac-5dc0-4319-92e8-203585c220c9): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:58:09 crc kubenswrapper[4767]: E0129 15:58:09.157509 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:58:20 crc kubenswrapper[4767]: E0129 15:58:20.020814 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:58:23 crc kubenswrapper[4767]: E0129 15:58:23.020645 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:58:32 crc kubenswrapper[4767]: E0129 15:58:32.162196 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:58:32 crc kubenswrapper[4767]: E0129 15:58:32.162886 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pvrz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5zmq2_openshift-marketplace(8b375cb3-2015-4d3f-b0d0-1e883fca2e5b): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:58:32 crc kubenswrapper[4767]: E0129 15:58:32.164199 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:58:38 crc kubenswrapper[4767]: E0129 15:58:38.163616 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:58:38 crc kubenswrapper[4767]: E0129 15:58:38.164513 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-w2zss_openshift-marketplace(5f7003ac-5dc0-4319-92e8-203585c220c9): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:58:38 crc kubenswrapper[4767]: E0129 15:58:38.166376 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:58:47 crc kubenswrapper[4767]: E0129 15:58:47.020880 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:58:52 crc kubenswrapper[4767]: E0129 15:58:52.020279 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:59:00 crc kubenswrapper[4767]: E0129 15:59:00.021238 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:59:06 crc kubenswrapper[4767]: E0129 15:59:06.022530 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:59:12 crc kubenswrapper[4767]: E0129 15:59:12.020821 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:59:18 crc kubenswrapper[4767]: E0129 15:59:18.021096 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:59:24 crc kubenswrapper[4767]: E0129 15:59:24.159279 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 29 15:59:24 crc kubenswrapper[4767]: E0129 15:59:24.160012 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pvrz8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5zmq2_openshift-marketplace(8b375cb3-2015-4d3f-b0d0-1e883fca2e5b): ErrImagePull: initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:59:24 crc kubenswrapper[4767]: E0129 15:59:24.161951 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/certified-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:59:31 crc kubenswrapper[4767]: E0129 15:59:31.212118 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 29 15:59:31 crc kubenswrapper[4767]: E0129 15:59:31.212753 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqnnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-w2zss_openshift-marketplace(5f7003ac-5dc0-4319-92e8-203585c220c9): ErrImagePull: initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 15:59:31 crc kubenswrapper[4767]: E0129 15:59:31.214154 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/community-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:59:37 crc kubenswrapper[4767]: E0129 15:59:37.021122 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:59:44 crc kubenswrapper[4767]: E0129 15:59:44.021814 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 15:59:50 crc kubenswrapper[4767]: E0129 15:59:50.020781 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 15:59:57 crc kubenswrapper[4767]: E0129 15:59:57.022667 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.144755 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl"] Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.145605 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.147972 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.148770 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.160018 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl"] Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.247459 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7070a452-2b24-4ce9-9bf9-287b80a5fa41-config-volume\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.247753 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjt56\" (UniqueName: \"kubernetes.io/projected/7070a452-2b24-4ce9-9bf9-287b80a5fa41-kube-api-access-bjt56\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.247780 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7070a452-2b24-4ce9-9bf9-287b80a5fa41-secret-volume\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.349843 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7070a452-2b24-4ce9-9bf9-287b80a5fa41-config-volume\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.349933 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjt56\" (UniqueName: \"kubernetes.io/projected/7070a452-2b24-4ce9-9bf9-287b80a5fa41-kube-api-access-bjt56\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.349964 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7070a452-2b24-4ce9-9bf9-287b80a5fa41-secret-volume\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.351145 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7070a452-2b24-4ce9-9bf9-287b80a5fa41-config-volume\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.356806 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7070a452-2b24-4ce9-9bf9-287b80a5fa41-secret-volume\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.367347 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjt56\" (UniqueName: \"kubernetes.io/projected/7070a452-2b24-4ce9-9bf9-287b80a5fa41-kube-api-access-bjt56\") pod \"collect-profiles-29495040-62lcl\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.479226 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:00 crc kubenswrapper[4767]: I0129 16:00:00.903873 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl"] Jan 29 16:00:01 crc kubenswrapper[4767]: I0129 16:00:01.536953 4767 generic.go:334] "Generic (PLEG): container finished" podID="7070a452-2b24-4ce9-9bf9-287b80a5fa41" containerID="70e6fb8d5543a3289e102b67745b85109c5130df0f630956398b0ca60db86ee2" exitCode=0 Jan 29 16:00:01 crc kubenswrapper[4767]: I0129 16:00:01.537020 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" event={"ID":"7070a452-2b24-4ce9-9bf9-287b80a5fa41","Type":"ContainerDied","Data":"70e6fb8d5543a3289e102b67745b85109c5130df0f630956398b0ca60db86ee2"} Jan 29 16:00:01 crc kubenswrapper[4767]: I0129 16:00:01.537243 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" event={"ID":"7070a452-2b24-4ce9-9bf9-287b80a5fa41","Type":"ContainerStarted","Data":"10c13fc1ddb04fc21125d9115c871d45ce59e1da42404ea503a386bff35c2830"} Jan 29 16:00:02 crc kubenswrapper[4767]: I0129 16:00:02.846485 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.008073 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7070a452-2b24-4ce9-9bf9-287b80a5fa41-secret-volume\") pod \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.008765 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7070a452-2b24-4ce9-9bf9-287b80a5fa41-config-volume\") pod \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.008815 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjt56\" (UniqueName: \"kubernetes.io/projected/7070a452-2b24-4ce9-9bf9-287b80a5fa41-kube-api-access-bjt56\") pod \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\" (UID: \"7070a452-2b24-4ce9-9bf9-287b80a5fa41\") " Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.010101 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7070a452-2b24-4ce9-9bf9-287b80a5fa41-config-volume" (OuterVolumeSpecName: "config-volume") pod "7070a452-2b24-4ce9-9bf9-287b80a5fa41" (UID: "7070a452-2b24-4ce9-9bf9-287b80a5fa41"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.015583 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7070a452-2b24-4ce9-9bf9-287b80a5fa41-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7070a452-2b24-4ce9-9bf9-287b80a5fa41" (UID: "7070a452-2b24-4ce9-9bf9-287b80a5fa41"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.015630 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7070a452-2b24-4ce9-9bf9-287b80a5fa41-kube-api-access-bjt56" (OuterVolumeSpecName: "kube-api-access-bjt56") pod "7070a452-2b24-4ce9-9bf9-287b80a5fa41" (UID: "7070a452-2b24-4ce9-9bf9-287b80a5fa41"). InnerVolumeSpecName "kube-api-access-bjt56". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:00:03 crc kubenswrapper[4767]: E0129 16:00:03.022754 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.110481 4767 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7070a452-2b24-4ce9-9bf9-287b80a5fa41-config-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.110520 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjt56\" (UniqueName: \"kubernetes.io/projected/7070a452-2b24-4ce9-9bf9-287b80a5fa41-kube-api-access-bjt56\") on node \"crc\" DevicePath \"\"" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.110546 4767 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7070a452-2b24-4ce9-9bf9-287b80a5fa41-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.555934 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" event={"ID":"7070a452-2b24-4ce9-9bf9-287b80a5fa41","Type":"ContainerDied","Data":"10c13fc1ddb04fc21125d9115c871d45ce59e1da42404ea503a386bff35c2830"} Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.555984 4767 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10c13fc1ddb04fc21125d9115c871d45ce59e1da42404ea503a386bff35c2830" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.556019 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29495040-62lcl" Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.919269 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt"] Jan 29 16:00:03 crc kubenswrapper[4767]: I0129 16:00:03.926446 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29494995-mcxqt"] Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.927278 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kfvzq/must-gather-lvfx5"] Jan 29 16:00:04 crc kubenswrapper[4767]: E0129 16:00:04.927928 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7070a452-2b24-4ce9-9bf9-287b80a5fa41" containerName="collect-profiles" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.927944 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="7070a452-2b24-4ce9-9bf9-287b80a5fa41" containerName="collect-profiles" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.928092 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="7070a452-2b24-4ce9-9bf9-287b80a5fa41" containerName="collect-profiles" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.928947 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.932349 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kfvzq"/"kube-root-ca.crt" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.932402 4767 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kfvzq"/"openshift-service-ca.crt" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.934147 4767 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-kfvzq"/"default-dockercfg-k76rp" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.936366 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lmnn\" (UniqueName: \"kubernetes.io/projected/5590f448-f43f-40c7-a902-672d79f0956a-kube-api-access-6lmnn\") pod \"must-gather-lvfx5\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.936431 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5590f448-f43f-40c7-a902-672d79f0956a-must-gather-output\") pod \"must-gather-lvfx5\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:04 crc kubenswrapper[4767]: I0129 16:00:04.943967 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kfvzq/must-gather-lvfx5"] Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.028870 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a83917d-1a8b-470e-8403-d28a2a74a81d" path="/var/lib/kubelet/pods/0a83917d-1a8b-470e-8403-d28a2a74a81d/volumes" Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.037559 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5590f448-f43f-40c7-a902-672d79f0956a-must-gather-output\") pod \"must-gather-lvfx5\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.037730 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lmnn\" (UniqueName: \"kubernetes.io/projected/5590f448-f43f-40c7-a902-672d79f0956a-kube-api-access-6lmnn\") pod \"must-gather-lvfx5\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.038147 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5590f448-f43f-40c7-a902-672d79f0956a-must-gather-output\") pod \"must-gather-lvfx5\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.088886 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lmnn\" (UniqueName: \"kubernetes.io/projected/5590f448-f43f-40c7-a902-672d79f0956a-kube-api-access-6lmnn\") pod \"must-gather-lvfx5\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.245726 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.481829 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kfvzq/must-gather-lvfx5"] Jan 29 16:00:05 crc kubenswrapper[4767]: I0129 16:00:05.570654 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" event={"ID":"5590f448-f43f-40c7-a902-672d79f0956a","Type":"ContainerStarted","Data":"9854e537ddb29eefb94c6e5b8f42a3a1bd96544a77274272e09ba2a285f92083"} Jan 29 16:00:09 crc kubenswrapper[4767]: E0129 16:00:09.977463 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 16:00:12 crc kubenswrapper[4767]: I0129 16:00:12.805801 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:00:12 crc kubenswrapper[4767]: I0129 16:00:12.805909 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:00:13 crc kubenswrapper[4767]: I0129 16:00:13.628130 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" event={"ID":"5590f448-f43f-40c7-a902-672d79f0956a","Type":"ContainerStarted","Data":"fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d"} Jan 29 16:00:13 crc kubenswrapper[4767]: I0129 16:00:13.628717 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" event={"ID":"5590f448-f43f-40c7-a902-672d79f0956a","Type":"ContainerStarted","Data":"ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd"} Jan 29 16:00:13 crc kubenswrapper[4767]: I0129 16:00:13.646627 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" podStartSLOduration=2.219607354 podStartE2EDuration="9.646606144s" podCreationTimestamp="2026-01-29 16:00:04 +0000 UTC" firstStartedPulling="2026-01-29 16:00:05.49301775 +0000 UTC m=+3561.303334789" lastFinishedPulling="2026-01-29 16:00:12.92001654 +0000 UTC m=+3568.730333579" observedRunningTime="2026-01-29 16:00:13.64612471 +0000 UTC m=+3569.456441749" watchObservedRunningTime="2026-01-29 16:00:13.646606144 +0000 UTC m=+3569.456923183" Jan 29 16:00:18 crc kubenswrapper[4767]: E0129 16:00:18.020785 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 16:00:23 crc kubenswrapper[4767]: E0129 16:00:23.020273 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 16:00:33 crc kubenswrapper[4767]: E0129 16:00:33.021060 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" Jan 29 16:00:35 crc kubenswrapper[4767]: E0129 16:00:35.024857 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 16:00:42 crc kubenswrapper[4767]: I0129 16:00:42.805589 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:00:42 crc kubenswrapper[4767]: I0129 16:00:42.806230 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:00:46 crc kubenswrapper[4767]: E0129 16:00:46.020426 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" Jan 29 16:00:46 crc kubenswrapper[4767]: I0129 16:00:46.407915 4767 scope.go:117] "RemoveContainer" containerID="7dcfed40d25e3788db44ed5e7cda52e787f83fb642353e85eb933f9d04268cbe" Jan 29 16:00:48 crc kubenswrapper[4767]: I0129 16:00:48.872040 4767 generic.go:334] "Generic (PLEG): container finished" podID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerID="1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b" exitCode=0 Jan 29 16:00:48 crc kubenswrapper[4767]: I0129 16:00:48.872152 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5zmq2" event={"ID":"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b","Type":"ContainerDied","Data":"1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b"} Jan 29 16:00:49 crc kubenswrapper[4767]: I0129 16:00:49.888961 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5zmq2" event={"ID":"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b","Type":"ContainerStarted","Data":"f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156"} Jan 29 16:00:49 crc kubenswrapper[4767]: I0129 16:00:49.920474 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5zmq2" podStartSLOduration=3.103588419 podStartE2EDuration="2m56.920456332s" podCreationTimestamp="2026-01-29 15:57:53 +0000 UTC" firstStartedPulling="2026-01-29 15:57:55.558054356 +0000 UTC m=+3431.368371405" lastFinishedPulling="2026-01-29 16:00:49.374922279 +0000 UTC m=+3605.185239318" observedRunningTime="2026-01-29 16:00:49.914534053 +0000 UTC m=+3605.724851102" watchObservedRunningTime="2026-01-29 16:00:49.920456332 +0000 UTC m=+3605.730773371" Jan 29 16:00:53 crc kubenswrapper[4767]: I0129 16:00:53.754844 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 16:00:53 crc kubenswrapper[4767]: I0129 16:00:53.755164 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 16:00:53 crc kubenswrapper[4767]: I0129 16:00:53.803707 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 16:00:56 crc kubenswrapper[4767]: I0129 16:00:56.945927 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f9wfq"] Jan 29 16:00:56 crc kubenswrapper[4767]: I0129 16:00:56.948926 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:56 crc kubenswrapper[4767]: I0129 16:00:56.952690 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9wfq"] Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.082824 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k74zw\" (UniqueName: \"kubernetes.io/projected/d220a90a-2e09-428f-899f-9a0abc241587-kube-api-access-k74zw\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.082971 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-catalog-content\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.083026 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-utilities\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.184717 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-utilities\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.184807 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k74zw\" (UniqueName: \"kubernetes.io/projected/d220a90a-2e09-428f-899f-9a0abc241587-kube-api-access-k74zw\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.184941 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-catalog-content\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.185693 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-catalog-content\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.185929 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-utilities\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.210801 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k74zw\" (UniqueName: \"kubernetes.io/projected/d220a90a-2e09-428f-899f-9a0abc241587-kube-api-access-k74zw\") pod \"redhat-marketplace-f9wfq\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:00:57 crc kubenswrapper[4767]: I0129 16:00:57.300698 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:01:00 crc kubenswrapper[4767]: I0129 16:00:57.750645 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9wfq"] Jan 29 16:01:00 crc kubenswrapper[4767]: W0129 16:00:57.755092 4767 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd220a90a_2e09_428f_899f_9a0abc241587.slice/crio-87aeb7de7e39885043c99dfdc5010de9066933ebfa1069a8267ce6cb1919ca03 WatchSource:0}: Error finding container 87aeb7de7e39885043c99dfdc5010de9066933ebfa1069a8267ce6cb1919ca03: Status 404 returned error can't find the container with id 87aeb7de7e39885043c99dfdc5010de9066933ebfa1069a8267ce6cb1919ca03 Jan 29 16:01:00 crc kubenswrapper[4767]: I0129 16:00:57.955471 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9wfq" event={"ID":"d220a90a-2e09-428f-899f-9a0abc241587","Type":"ContainerStarted","Data":"87aeb7de7e39885043c99dfdc5010de9066933ebfa1069a8267ce6cb1919ca03"} Jan 29 16:01:00 crc kubenswrapper[4767]: I0129 16:00:58.963722 4767 generic.go:334] "Generic (PLEG): container finished" podID="d220a90a-2e09-428f-899f-9a0abc241587" containerID="f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b" exitCode=0 Jan 29 16:01:00 crc kubenswrapper[4767]: I0129 16:00:58.963970 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9wfq" event={"ID":"d220a90a-2e09-428f-899f-9a0abc241587","Type":"ContainerDied","Data":"f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b"} Jan 29 16:01:00 crc kubenswrapper[4767]: I0129 16:01:00.982238 4767 generic.go:334] "Generic (PLEG): container finished" podID="d220a90a-2e09-428f-899f-9a0abc241587" containerID="100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0" exitCode=0 Jan 29 16:01:00 crc kubenswrapper[4767]: I0129 16:01:00.982321 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9wfq" event={"ID":"d220a90a-2e09-428f-899f-9a0abc241587","Type":"ContainerDied","Data":"100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0"} Jan 29 16:01:00 crc kubenswrapper[4767]: I0129 16:01:00.984780 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zss" event={"ID":"5f7003ac-5dc0-4319-92e8-203585c220c9","Type":"ContainerStarted","Data":"b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1"} Jan 29 16:01:01 crc kubenswrapper[4767]: I0129 16:01:01.995467 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9wfq" event={"ID":"d220a90a-2e09-428f-899f-9a0abc241587","Type":"ContainerStarted","Data":"4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73"} Jan 29 16:01:01 crc kubenswrapper[4767]: I0129 16:01:01.998016 4767 generic.go:334] "Generic (PLEG): container finished" podID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerID="b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1" exitCode=0 Jan 29 16:01:01 crc kubenswrapper[4767]: I0129 16:01:01.998047 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zss" event={"ID":"5f7003ac-5dc0-4319-92e8-203585c220c9","Type":"ContainerDied","Data":"b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1"} Jan 29 16:01:02 crc kubenswrapper[4767]: I0129 16:01:02.017777 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f9wfq" podStartSLOduration=3.543795502 podStartE2EDuration="6.017756124s" podCreationTimestamp="2026-01-29 16:00:56 +0000 UTC" firstStartedPulling="2026-01-29 16:00:58.965583461 +0000 UTC m=+3614.775900500" lastFinishedPulling="2026-01-29 16:01:01.439544083 +0000 UTC m=+3617.249861122" observedRunningTime="2026-01-29 16:01:02.012262468 +0000 UTC m=+3617.822579507" watchObservedRunningTime="2026-01-29 16:01:02.017756124 +0000 UTC m=+3617.828073173" Jan 29 16:01:03 crc kubenswrapper[4767]: I0129 16:01:03.006718 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zss" event={"ID":"5f7003ac-5dc0-4319-92e8-203585c220c9","Type":"ContainerStarted","Data":"f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd"} Jan 29 16:01:03 crc kubenswrapper[4767]: I0129 16:01:03.032124 4767 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w2zss" podStartSLOduration=2.797772158 podStartE2EDuration="3m13.032105836s" podCreationTimestamp="2026-01-29 15:57:50 +0000 UTC" firstStartedPulling="2026-01-29 15:57:52.523100893 +0000 UTC m=+3428.333417942" lastFinishedPulling="2026-01-29 16:01:02.757434581 +0000 UTC m=+3618.567751620" observedRunningTime="2026-01-29 16:01:03.025767876 +0000 UTC m=+3618.836084915" watchObservedRunningTime="2026-01-29 16:01:03.032105836 +0000 UTC m=+3618.842422875" Jan 29 16:01:03 crc kubenswrapper[4767]: I0129 16:01:03.801218 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.301884 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.302405 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.329766 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5zmq2"] Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.330020 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5zmq2" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="registry-server" containerID="cri-o://f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156" gracePeriod=2 Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.348290 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.749092 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.930369 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-utilities\") pod \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.930446 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-catalog-content\") pod \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.930479 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvrz8\" (UniqueName: \"kubernetes.io/projected/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-kube-api-access-pvrz8\") pod \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\" (UID: \"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b\") " Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.931352 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-utilities" (OuterVolumeSpecName: "utilities") pod "8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" (UID: "8b375cb3-2015-4d3f-b0d0-1e883fca2e5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.936549 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-kube-api-access-pvrz8" (OuterVolumeSpecName: "kube-api-access-pvrz8") pod "8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" (UID: "8b375cb3-2015-4d3f-b0d0-1e883fca2e5b"). InnerVolumeSpecName "kube-api-access-pvrz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:01:07 crc kubenswrapper[4767]: I0129 16:01:07.975971 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" (UID: "8b375cb3-2015-4d3f-b0d0-1e883fca2e5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.032416 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.032455 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.032470 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvrz8\" (UniqueName: \"kubernetes.io/projected/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b-kube-api-access-pvrz8\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.052132 4767 generic.go:334] "Generic (PLEG): container finished" podID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerID="f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156" exitCode=0 Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.052179 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5zmq2" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.052199 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5zmq2" event={"ID":"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b","Type":"ContainerDied","Data":"f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156"} Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.052249 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5zmq2" event={"ID":"8b375cb3-2015-4d3f-b0d0-1e883fca2e5b","Type":"ContainerDied","Data":"50a2c587606b0ef14bd0b500be2a0a06e62acb3501f4bf04020c241f42cbb57e"} Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.052270 4767 scope.go:117] "RemoveContainer" containerID="f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.076186 4767 scope.go:117] "RemoveContainer" containerID="1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.083975 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5zmq2"] Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.091225 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5zmq2"] Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.109115 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.110569 4767 scope.go:117] "RemoveContainer" containerID="c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.133241 4767 scope.go:117] "RemoveContainer" containerID="f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156" Jan 29 16:01:08 crc kubenswrapper[4767]: E0129 16:01:08.133715 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156\": container with ID starting with f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156 not found: ID does not exist" containerID="f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.133764 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156"} err="failed to get container status \"f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156\": rpc error: code = NotFound desc = could not find container \"f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156\": container with ID starting with f0d1de5b99db9f3bf08802d54014774a73d439059109cb5c01e19c504e518156 not found: ID does not exist" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.133795 4767 scope.go:117] "RemoveContainer" containerID="1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b" Jan 29 16:01:08 crc kubenswrapper[4767]: E0129 16:01:08.134181 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b\": container with ID starting with 1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b not found: ID does not exist" containerID="1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.134217 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b"} err="failed to get container status \"1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b\": rpc error: code = NotFound desc = could not find container \"1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b\": container with ID starting with 1f84a880c594e7384dbc903607f2574883255731b81050716cce7f2ef6a31a6b not found: ID does not exist" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.134235 4767 scope.go:117] "RemoveContainer" containerID="c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3" Jan 29 16:01:08 crc kubenswrapper[4767]: E0129 16:01:08.134644 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3\": container with ID starting with c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3 not found: ID does not exist" containerID="c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3" Jan 29 16:01:08 crc kubenswrapper[4767]: I0129 16:01:08.134719 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3"} err="failed to get container status \"c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3\": rpc error: code = NotFound desc = could not find container \"c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3\": container with ID starting with c14b5dcee0e049bfb2c7bf4172bb0242422162c2d121a7736db2fc03bd0bb8e3 not found: ID does not exist" Jan 29 16:01:09 crc kubenswrapper[4767]: I0129 16:01:09.028923 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" path="/var/lib/kubelet/pods/8b375cb3-2015-4d3f-b0d0-1e883fca2e5b/volumes" Jan 29 16:01:09 crc kubenswrapper[4767]: I0129 16:01:09.728885 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9wfq"] Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.067516 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f9wfq" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="registry-server" containerID="cri-o://4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73" gracePeriod=2 Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.509331 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.585858 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm_ee2c1e3b-70b5-4479-a3f2-aa33aba08243/util/0.log" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.667706 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-utilities\") pod \"d220a90a-2e09-428f-899f-9a0abc241587\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.667782 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-catalog-content\") pod \"d220a90a-2e09-428f-899f-9a0abc241587\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.667854 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k74zw\" (UniqueName: \"kubernetes.io/projected/d220a90a-2e09-428f-899f-9a0abc241587-kube-api-access-k74zw\") pod \"d220a90a-2e09-428f-899f-9a0abc241587\" (UID: \"d220a90a-2e09-428f-899f-9a0abc241587\") " Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.669605 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-utilities" (OuterVolumeSpecName: "utilities") pod "d220a90a-2e09-428f-899f-9a0abc241587" (UID: "d220a90a-2e09-428f-899f-9a0abc241587"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.692114 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d220a90a-2e09-428f-899f-9a0abc241587-kube-api-access-k74zw" (OuterVolumeSpecName: "kube-api-access-k74zw") pod "d220a90a-2e09-428f-899f-9a0abc241587" (UID: "d220a90a-2e09-428f-899f-9a0abc241587"). InnerVolumeSpecName "kube-api-access-k74zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.703015 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d220a90a-2e09-428f-899f-9a0abc241587" (UID: "d220a90a-2e09-428f-899f-9a0abc241587"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.769976 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k74zw\" (UniqueName: \"kubernetes.io/projected/d220a90a-2e09-428f-899f-9a0abc241587-kube-api-access-k74zw\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.770037 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.770054 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d220a90a-2e09-428f-899f-9a0abc241587-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:10 crc kubenswrapper[4767]: I0129 16:01:10.986251 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm_ee2c1e3b-70b5-4479-a3f2-aa33aba08243/util/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.000654 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm_ee2c1e3b-70b5-4479-a3f2-aa33aba08243/pull/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.046291 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm_ee2c1e3b-70b5-4479-a3f2-aa33aba08243/pull/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.065581 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm_ee2c1e3b-70b5-4479-a3f2-aa33aba08243/util/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.074956 4767 generic.go:334] "Generic (PLEG): container finished" podID="d220a90a-2e09-428f-899f-9a0abc241587" containerID="4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73" exitCode=0 Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.075000 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9wfq" event={"ID":"d220a90a-2e09-428f-899f-9a0abc241587","Type":"ContainerDied","Data":"4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73"} Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.075032 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f9wfq" event={"ID":"d220a90a-2e09-428f-899f-9a0abc241587","Type":"ContainerDied","Data":"87aeb7de7e39885043c99dfdc5010de9066933ebfa1069a8267ce6cb1919ca03"} Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.075054 4767 scope.go:117] "RemoveContainer" containerID="4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.075219 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f9wfq" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.099985 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9wfq"] Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.100675 4767 scope.go:117] "RemoveContainer" containerID="100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.105932 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f9wfq"] Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.120110 4767 scope.go:117] "RemoveContainer" containerID="f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.150502 4767 scope.go:117] "RemoveContainer" containerID="4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73" Jan 29 16:01:11 crc kubenswrapper[4767]: E0129 16:01:11.150963 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73\": container with ID starting with 4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73 not found: ID does not exist" containerID="4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.150993 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73"} err="failed to get container status \"4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73\": rpc error: code = NotFound desc = could not find container \"4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73\": container with ID starting with 4bb04d3ec1df64a8a29bc1c8ff9eac8aadc7ca2f554ce6359bed0fc807d29a73 not found: ID does not exist" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.151016 4767 scope.go:117] "RemoveContainer" containerID="100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0" Jan 29 16:01:11 crc kubenswrapper[4767]: E0129 16:01:11.151243 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0\": container with ID starting with 100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0 not found: ID does not exist" containerID="100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.151261 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0"} err="failed to get container status \"100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0\": rpc error: code = NotFound desc = could not find container \"100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0\": container with ID starting with 100cebef284abf4c07e25f73469d84b0ef2c497f82265bccf8011059841d8ec0 not found: ID does not exist" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.151272 4767 scope.go:117] "RemoveContainer" containerID="f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b" Jan 29 16:01:11 crc kubenswrapper[4767]: E0129 16:01:11.151442 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b\": container with ID starting with f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b not found: ID does not exist" containerID="f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.151458 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b"} err="failed to get container status \"f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b\": rpc error: code = NotFound desc = could not find container \"f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b\": container with ID starting with f1ff14d40d9ccdaccfc7edb49c36e8c5e300db8ecb0a69420cb8c48bace5e86b not found: ID does not exist" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.184717 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm_ee2c1e3b-70b5-4479-a3f2-aa33aba08243/pull/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.240949 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_278bb120acdf091c21190c1c46edbfd3b5935fa79682fd6780673405cemrrnm_ee2c1e3b-70b5-4479-a3f2-aa33aba08243/extract/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.318131 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d6fdb96dc-gkq4d_157b230c-4bbc-4a62-841e-60acba588f1e/manager/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.450559 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-858d89fd-v5bpj_4be7d2d9-0fa7-4783-987a-7dd06823b118/manager/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.574229 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-dd77988f8-bvxsd_49292974-81e4-4eba-8dc8-d6ac21111b9a/manager/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.637581 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-f8c4db9df-r4nvx_440d7f0e-a462-453a-b41f-ccdb1521bb20/manager/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.774581 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-d8b84fbc-9zkzh_a1b7dc5f-dccb-44e8-b390-daa6f9e773b3/manager/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.842771 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w2zss" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.842830 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w2zss" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.857348 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5fb775575f-6gtvq_a53807b1-819e-4e75-9f4b-ebda7fc7b415/manager/0.log" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.894061 4767 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w2zss" Jan 29 16:01:11 crc kubenswrapper[4767]: I0129 16:01:11.956969 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79955696d6-br5bs_97c4098d-defd-4156-b48b-e26dd157f709/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.117581 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-866c9d5b98-hvtlx_68ca9f6a-ab9f-4188-a45c-23e6c1303a20/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.128845 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-8ccc8547b-gd2cz_6e272311-00bb-4cfb-92b5-2cf77eb483f8/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.140960 4767 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w2zss" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.305575 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-76c896469f-xxrwp_eba759c2-d8c7-4956-a47c-def368a54bf1/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.334600 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67bf948998-rlgl7_3350030b-3cea-4a51-9fdd-34408ad506eb/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.511116 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-68cb478976-gbtpt_833297bf-cdda-441d-8e03-de62a3b0c5b3/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.515648 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c7cc6ff45-zwfl4_a7f3205f-40e9-4eed-8305-4e0ca0ebe75c/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.665664 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-68f8cb846c-cs4g8_ca1f1109-544f-490f-9bf5-9d05241d30f2/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.735828 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-59c4b45c4dr2pn4_dc435770-6862-4056-898e-ee9fbf4850d3/manager/0.log" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.806218 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.806286 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.806338 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.807017 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"55581c47200bca1e0d0702088fd9033cee9ab992eba215fd248062d21ef80f24"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.807087 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://55581c47200bca1e0d0702088fd9033cee9ab992eba215fd248062d21ef80f24" gracePeriod=600 Jan 29 16:01:12 crc kubenswrapper[4767]: I0129 16:01:12.989270 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-68476f6fbd-2l847_bf5fbc7f-b1a6-4a4a-a4e5-6f66e583e25e/operator/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.030269 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d220a90a-2e09-428f-899f-9a0abc241587" path="/var/lib/kubelet/pods/d220a90a-2e09-428f-899f-9a0abc241587/volumes" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.034717 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54c8dbf95-blw98_b2a9e907-bdb8-408f-ba23-b13a5f792e5a/manager/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.160843 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tsfnc_12bdee23-d478-434a-95bc-7d2c042f6906/registry-server/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.200879 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-788c46999f-gbkxz_0d956603-d824-432d-b1db-6574d2bde092/manager/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.391201 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5b964cf4cd-fsgxh_1a51e166-62bf-4181-b44f-62d52b543f2e/manager/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.503800 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-xmjrr_f104967e-4b06-42b6-b9ea-c4edb1923e87/operator/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.644011 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6f7455757b-pc4p9_adc494cd-a7d7-4104-9d5b-165115dfe504/manager/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.735980 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2zss"] Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.758785 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cf8c44c7-nzz8p_b186845a-5b3b-42ce-aad0-0269580f5b07/manager/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.834235 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56f8bfcd9f-dbphp_7b712b0e-3913-4421-a6c8-90991284909a/manager/0.log" Jan 29 16:01:13 crc kubenswrapper[4767]: I0129 16:01:13.931927 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59f4c7d7c4-g6hsv_e63115ec-b52a-4e9f-89b5-76004e0f6316/manager/0.log" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.101462 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="55581c47200bca1e0d0702088fd9033cee9ab992eba215fd248062d21ef80f24" exitCode=0 Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.101546 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"55581c47200bca1e0d0702088fd9033cee9ab992eba215fd248062d21ef80f24"} Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.101656 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w2zss" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="registry-server" containerID="cri-o://f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd" gracePeriod=2 Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.101647 4767 scope.go:117] "RemoveContainer" containerID="2f3f85425d105c2e3362115e9c6cc936116293f10082cbf2f0f43a14ab12bb7e" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.578031 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zss" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.725079 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-utilities\") pod \"5f7003ac-5dc0-4319-92e8-203585c220c9\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.725378 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-catalog-content\") pod \"5f7003ac-5dc0-4319-92e8-203585c220c9\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.725454 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqnnp\" (UniqueName: \"kubernetes.io/projected/5f7003ac-5dc0-4319-92e8-203585c220c9-kube-api-access-bqnnp\") pod \"5f7003ac-5dc0-4319-92e8-203585c220c9\" (UID: \"5f7003ac-5dc0-4319-92e8-203585c220c9\") " Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.726031 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-utilities" (OuterVolumeSpecName: "utilities") pod "5f7003ac-5dc0-4319-92e8-203585c220c9" (UID: "5f7003ac-5dc0-4319-92e8-203585c220c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.742076 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f7003ac-5dc0-4319-92e8-203585c220c9-kube-api-access-bqnnp" (OuterVolumeSpecName: "kube-api-access-bqnnp") pod "5f7003ac-5dc0-4319-92e8-203585c220c9" (UID: "5f7003ac-5dc0-4319-92e8-203585c220c9"). InnerVolumeSpecName "kube-api-access-bqnnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.790474 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f7003ac-5dc0-4319-92e8-203585c220c9" (UID: "5f7003ac-5dc0-4319-92e8-203585c220c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.826828 4767 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-utilities\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.826860 4767 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f7003ac-5dc0-4319-92e8-203585c220c9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:14 crc kubenswrapper[4767]: I0129 16:01:14.826871 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqnnp\" (UniqueName: \"kubernetes.io/projected/5f7003ac-5dc0-4319-92e8-203585c220c9-kube-api-access-bqnnp\") on node \"crc\" DevicePath \"\"" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.109524 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerStarted","Data":"2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2"} Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.113028 4767 generic.go:334] "Generic (PLEG): container finished" podID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerID="f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd" exitCode=0 Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.113056 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zss" event={"ID":"5f7003ac-5dc0-4319-92e8-203585c220c9","Type":"ContainerDied","Data":"f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd"} Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.113074 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w2zss" event={"ID":"5f7003ac-5dc0-4319-92e8-203585c220c9","Type":"ContainerDied","Data":"a3af4104f25b5f9acc1941219dcd69af57a64bf15445a696b7b70db330805c12"} Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.113089 4767 scope.go:117] "RemoveContainer" containerID="f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.113205 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w2zss" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.133329 4767 scope.go:117] "RemoveContainer" containerID="b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.157344 4767 scope.go:117] "RemoveContainer" containerID="6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.160981 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w2zss"] Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.167097 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w2zss"] Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.187740 4767 scope.go:117] "RemoveContainer" containerID="f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd" Jan 29 16:01:15 crc kubenswrapper[4767]: E0129 16:01:15.189043 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd\": container with ID starting with f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd not found: ID does not exist" containerID="f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.189087 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd"} err="failed to get container status \"f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd\": rpc error: code = NotFound desc = could not find container \"f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd\": container with ID starting with f26948cc3bd528d2868e2c95e1443f88afb1d1c5523da5fab7657040c0bd2fdd not found: ID does not exist" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.189117 4767 scope.go:117] "RemoveContainer" containerID="b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1" Jan 29 16:01:15 crc kubenswrapper[4767]: E0129 16:01:15.189436 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1\": container with ID starting with b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1 not found: ID does not exist" containerID="b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.189481 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1"} err="failed to get container status \"b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1\": rpc error: code = NotFound desc = could not find container \"b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1\": container with ID starting with b4c7da1ab6ba0648f04851372c72d24bc01d5edc535bdd7b534df68048cea3c1 not found: ID does not exist" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.189510 4767 scope.go:117] "RemoveContainer" containerID="6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c" Jan 29 16:01:15 crc kubenswrapper[4767]: E0129 16:01:15.191478 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c\": container with ID starting with 6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c not found: ID does not exist" containerID="6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c" Jan 29 16:01:15 crc kubenswrapper[4767]: I0129 16:01:15.191525 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c"} err="failed to get container status \"6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c\": rpc error: code = NotFound desc = could not find container \"6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c\": container with ID starting with 6a7820dc747773b022fb94c689f5088a1cc5d0e9c20cfa87064ffbb48967335c not found: ID does not exist" Jan 29 16:01:17 crc kubenswrapper[4767]: I0129 16:01:17.025975 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" path="/var/lib/kubelet/pods/5f7003ac-5dc0-4319-92e8-203585c220c9/volumes" Jan 29 16:01:32 crc kubenswrapper[4767]: I0129 16:01:32.968741 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-qxdzw_b8a5533d-d711-4fa8-ab7a-a686f77cb578/control-plane-machine-set-operator/0.log" Jan 29 16:01:33 crc kubenswrapper[4767]: I0129 16:01:33.172715 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q9hsl_e3cbd296-0674-49b6-b9a5-7bb9f6736387/kube-rbac-proxy/0.log" Jan 29 16:01:33 crc kubenswrapper[4767]: I0129 16:01:33.210834 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q9hsl_e3cbd296-0674-49b6-b9a5-7bb9f6736387/machine-api-operator/0.log" Jan 29 16:01:45 crc kubenswrapper[4767]: I0129 16:01:45.367159 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-tztj2_e9862466-1c30-4644-bf55-8e9f1548555e/cert-manager-controller/0.log" Jan 29 16:01:45 crc kubenswrapper[4767]: I0129 16:01:45.561576 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-k46vw_66377cb0-a5db-4193-888c-8252e286f1a4/cert-manager-cainjector/0.log" Jan 29 16:01:45 crc kubenswrapper[4767]: I0129 16:01:45.680618 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-hp7hd_0649463b-73be-45e2-a8f3-393f0003d054/cert-manager-webhook/0.log" Jan 29 16:01:57 crc kubenswrapper[4767]: I0129 16:01:57.145636 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-mrsvg_cdbac033-2b2b-4f8a-a738-7304d0b0ac41/nmstate-console-plugin/0.log" Jan 29 16:01:57 crc kubenswrapper[4767]: I0129 16:01:57.277343 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-7pnbq_bc27129c-3bf6-4aeb-8a26-b46a555fe71f/nmstate-handler/0.log" Jan 29 16:01:57 crc kubenswrapper[4767]: I0129 16:01:57.312163 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-5dlzw_59099fce-32f9-4fc6-8ce4-e7789c00f336/kube-rbac-proxy/0.log" Jan 29 16:01:57 crc kubenswrapper[4767]: I0129 16:01:57.363499 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-5dlzw_59099fce-32f9-4fc6-8ce4-e7789c00f336/nmstate-metrics/0.log" Jan 29 16:01:57 crc kubenswrapper[4767]: I0129 16:01:57.530156 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-p4bhf_6bcf1dba-60e6-4d8e-adb1-cd00eb9309ea/nmstate-webhook/0.log" Jan 29 16:01:57 crc kubenswrapper[4767]: I0129 16:01:57.533187 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-9bjzl_1a87ddd0-f27d-4331-9e42-b13d173825bb/nmstate-operator/0.log" Jan 29 16:02:23 crc kubenswrapper[4767]: I0129 16:02:23.850674 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-h5qcl_dd2bff9a-d96a-4e6d-b8b6-517b692c2c44/kube-rbac-proxy/0.log" Jan 29 16:02:23 crc kubenswrapper[4767]: I0129 16:02:23.939787 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-h5qcl_dd2bff9a-d96a-4e6d-b8b6-517b692c2c44/controller/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.074855 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-frr-files/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.254203 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-metrics/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.261986 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-reloader/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.283879 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-reloader/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.289384 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-frr-files/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.482455 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-metrics/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.492528 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-reloader/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.505188 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-metrics/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.510374 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-frr-files/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.651999 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-reloader/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.670836 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-metrics/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.674112 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/cp-frr-files/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.713462 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/controller/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.858557 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/frr-metrics/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.865636 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/kube-rbac-proxy/0.log" Jan 29 16:02:24 crc kubenswrapper[4767]: I0129 16:02:24.906632 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/kube-rbac-proxy-frr/0.log" Jan 29 16:02:25 crc kubenswrapper[4767]: I0129 16:02:25.068663 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/reloader/0.log" Jan 29 16:02:25 crc kubenswrapper[4767]: I0129 16:02:25.096419 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-6b5vt_4304ecf7-d38c-4593-bd30-7e8de36c7638/frr/0.log" Jan 29 16:02:25 crc kubenswrapper[4767]: I0129 16:02:25.151178 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-qcf7k_1e14bfd9-9551-45e2-85cd-75f27bc10ada/frr-k8s-webhook-server/0.log" Jan 29 16:02:25 crc kubenswrapper[4767]: I0129 16:02:25.267914 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-875d98f97-bhdtk_80da1038-f75e-4ed8-9b53-dbd69e825d5e/manager/0.log" Jan 29 16:02:25 crc kubenswrapper[4767]: I0129 16:02:25.332851 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-589648c7bd-vm5bb_32caa2be-0b0c-4e06-b281-4f405ff743b6/webhook-server/0.log" Jan 29 16:02:25 crc kubenswrapper[4767]: I0129 16:02:25.478673 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ldqv6_6ba067f0-0799-4fb8-ba5c-20745eaced80/kube-rbac-proxy/0.log" Jan 29 16:02:25 crc kubenswrapper[4767]: I0129 16:02:25.600414 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-ldqv6_6ba067f0-0799-4fb8-ba5c-20745eaced80/speaker/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.259158 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg_f5ab91f4-430d-45e9-b662-fc611e22e033/util/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.399786 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg_f5ab91f4-430d-45e9-b662-fc611e22e033/util/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.450721 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg_f5ab91f4-430d-45e9-b662-fc611e22e033/pull/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.451006 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg_f5ab91f4-430d-45e9-b662-fc611e22e033/pull/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.623021 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg_f5ab91f4-430d-45e9-b662-fc611e22e033/util/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.666140 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg_f5ab91f4-430d-45e9-b662-fc611e22e033/pull/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.671365 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcq24qg_f5ab91f4-430d-45e9-b662-fc611e22e033/extract/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.817962 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g_c36ab79e-316a-41b0-94fa-387a7fa88d5c/util/0.log" Jan 29 16:02:38 crc kubenswrapper[4767]: I0129 16:02:38.996807 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g_c36ab79e-316a-41b0-94fa-387a7fa88d5c/pull/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.008946 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g_c36ab79e-316a-41b0-94fa-387a7fa88d5c/util/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.021675 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g_c36ab79e-316a-41b0-94fa-387a7fa88d5c/pull/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.185205 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g_c36ab79e-316a-41b0-94fa-387a7fa88d5c/pull/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.219877 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g_c36ab79e-316a-41b0-94fa-387a7fa88d5c/extract/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.223508 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713zqp9g_c36ab79e-316a-41b0-94fa-387a7fa88d5c/util/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.388380 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7_e3d9aea9-f021-4d1a-866c-4ef98a702bae/util/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.500287 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7_e3d9aea9-f021-4d1a-866c-4ef98a702bae/util/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.526110 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7_e3d9aea9-f021-4d1a-866c-4ef98a702bae/pull/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.570131 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7_e3d9aea9-f021-4d1a-866c-4ef98a702bae/pull/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.731429 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7_e3d9aea9-f021-4d1a-866c-4ef98a702bae/pull/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.743731 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7_e3d9aea9-f021-4d1a-866c-4ef98a702bae/extract/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.790145 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5ghlq7_e3d9aea9-f021-4d1a-866c-4ef98a702bae/util/0.log" Jan 29 16:02:39 crc kubenswrapper[4767]: I0129 16:02:39.913518 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zn5dm_149720b9-74d1-432e-ad66-7824a18ff8a5/extract-utilities/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.220339 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zn5dm_149720b9-74d1-432e-ad66-7824a18ff8a5/extract-utilities/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.225850 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zn5dm_149720b9-74d1-432e-ad66-7824a18ff8a5/extract-content/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.310143 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zn5dm_149720b9-74d1-432e-ad66-7824a18ff8a5/extract-content/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.444819 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zn5dm_149720b9-74d1-432e-ad66-7824a18ff8a5/extract-utilities/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.477610 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zn5dm_149720b9-74d1-432e-ad66-7824a18ff8a5/extract-content/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.669304 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tst6c_e1ef69eb-5928-4466-988d-6d4955721b2e/extract-utilities/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.838780 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zn5dm_149720b9-74d1-432e-ad66-7824a18ff8a5/registry-server/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.860588 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tst6c_e1ef69eb-5928-4466-988d-6d4955721b2e/extract-utilities/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.898403 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tst6c_e1ef69eb-5928-4466-988d-6d4955721b2e/extract-content/0.log" Jan 29 16:02:40 crc kubenswrapper[4767]: I0129 16:02:40.941554 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tst6c_e1ef69eb-5928-4466-988d-6d4955721b2e/extract-content/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.091334 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tst6c_e1ef69eb-5928-4466-988d-6d4955721b2e/extract-content/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.108320 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tst6c_e1ef69eb-5928-4466-988d-6d4955721b2e/extract-utilities/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.328030 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-g7t6n_dc351b10-105e-46dd-bef5-b8a19abaad73/marketplace-operator/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.416190 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c9h8z_3ff6370b-64ad-49cf-8f74-eb50bb30d826/extract-utilities/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.523824 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tst6c_e1ef69eb-5928-4466-988d-6d4955721b2e/registry-server/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.682345 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c9h8z_3ff6370b-64ad-49cf-8f74-eb50bb30d826/extract-utilities/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.691641 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c9h8z_3ff6370b-64ad-49cf-8f74-eb50bb30d826/extract-content/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.726554 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c9h8z_3ff6370b-64ad-49cf-8f74-eb50bb30d826/extract-content/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.846175 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c9h8z_3ff6370b-64ad-49cf-8f74-eb50bb30d826/extract-utilities/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.868993 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c9h8z_3ff6370b-64ad-49cf-8f74-eb50bb30d826/extract-content/0.log" Jan 29 16:02:41 crc kubenswrapper[4767]: I0129 16:02:41.929196 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c9h8z_3ff6370b-64ad-49cf-8f74-eb50bb30d826/registry-server/0.log" Jan 29 16:02:42 crc kubenswrapper[4767]: I0129 16:02:42.063861 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sp2xf_e635388d-4adc-4dfe-aca6-4ed3c7ecc646/extract-utilities/0.log" Jan 29 16:02:42 crc kubenswrapper[4767]: I0129 16:02:42.221532 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sp2xf_e635388d-4adc-4dfe-aca6-4ed3c7ecc646/extract-utilities/0.log" Jan 29 16:02:42 crc kubenswrapper[4767]: I0129 16:02:42.244325 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sp2xf_e635388d-4adc-4dfe-aca6-4ed3c7ecc646/extract-content/0.log" Jan 29 16:02:42 crc kubenswrapper[4767]: I0129 16:02:42.263562 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sp2xf_e635388d-4adc-4dfe-aca6-4ed3c7ecc646/extract-content/0.log" Jan 29 16:02:42 crc kubenswrapper[4767]: I0129 16:02:42.465712 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sp2xf_e635388d-4adc-4dfe-aca6-4ed3c7ecc646/extract-utilities/0.log" Jan 29 16:02:42 crc kubenswrapper[4767]: I0129 16:02:42.465828 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sp2xf_e635388d-4adc-4dfe-aca6-4ed3c7ecc646/extract-content/0.log" Jan 29 16:02:42 crc kubenswrapper[4767]: I0129 16:02:42.792783 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sp2xf_e635388d-4adc-4dfe-aca6-4ed3c7ecc646/registry-server/0.log" Jan 29 16:03:42 crc kubenswrapper[4767]: I0129 16:03:42.805320 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:03:42 crc kubenswrapper[4767]: I0129 16:03:42.805863 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:03:54 crc kubenswrapper[4767]: I0129 16:03:54.252942 4767 generic.go:334] "Generic (PLEG): container finished" podID="5590f448-f43f-40c7-a902-672d79f0956a" containerID="ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd" exitCode=0 Jan 29 16:03:54 crc kubenswrapper[4767]: I0129 16:03:54.253017 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" event={"ID":"5590f448-f43f-40c7-a902-672d79f0956a","Type":"ContainerDied","Data":"ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd"} Jan 29 16:03:54 crc kubenswrapper[4767]: I0129 16:03:54.254017 4767 scope.go:117] "RemoveContainer" containerID="ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd" Jan 29 16:03:54 crc kubenswrapper[4767]: I0129 16:03:54.909794 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kfvzq_must-gather-lvfx5_5590f448-f43f-40c7-a902-672d79f0956a/gather/0.log" Jan 29 16:04:02 crc kubenswrapper[4767]: I0129 16:04:02.440078 4767 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kfvzq/must-gather-lvfx5"] Jan 29 16:04:02 crc kubenswrapper[4767]: I0129 16:04:02.441010 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" podUID="5590f448-f43f-40c7-a902-672d79f0956a" containerName="copy" containerID="cri-o://fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d" gracePeriod=2 Jan 29 16:04:02 crc kubenswrapper[4767]: I0129 16:04:02.451190 4767 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kfvzq/must-gather-lvfx5"] Jan 29 16:04:02 crc kubenswrapper[4767]: I0129 16:04:02.915451 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kfvzq_must-gather-lvfx5_5590f448-f43f-40c7-a902-672d79f0956a/copy/0.log" Jan 29 16:04:02 crc kubenswrapper[4767]: I0129 16:04:02.916775 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.039143 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lmnn\" (UniqueName: \"kubernetes.io/projected/5590f448-f43f-40c7-a902-672d79f0956a-kube-api-access-6lmnn\") pod \"5590f448-f43f-40c7-a902-672d79f0956a\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.040101 4767 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5590f448-f43f-40c7-a902-672d79f0956a-must-gather-output\") pod \"5590f448-f43f-40c7-a902-672d79f0956a\" (UID: \"5590f448-f43f-40c7-a902-672d79f0956a\") " Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.045268 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5590f448-f43f-40c7-a902-672d79f0956a-kube-api-access-6lmnn" (OuterVolumeSpecName: "kube-api-access-6lmnn") pod "5590f448-f43f-40c7-a902-672d79f0956a" (UID: "5590f448-f43f-40c7-a902-672d79f0956a"). InnerVolumeSpecName "kube-api-access-6lmnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.125558 4767 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5590f448-f43f-40c7-a902-672d79f0956a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "5590f448-f43f-40c7-a902-672d79f0956a" (UID: "5590f448-f43f-40c7-a902-672d79f0956a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.142339 4767 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lmnn\" (UniqueName: \"kubernetes.io/projected/5590f448-f43f-40c7-a902-672d79f0956a-kube-api-access-6lmnn\") on node \"crc\" DevicePath \"\"" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.142384 4767 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/5590f448-f43f-40c7-a902-672d79f0956a-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.320966 4767 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kfvzq_must-gather-lvfx5_5590f448-f43f-40c7-a902-672d79f0956a/copy/0.log" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.321359 4767 generic.go:334] "Generic (PLEG): container finished" podID="5590f448-f43f-40c7-a902-672d79f0956a" containerID="fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d" exitCode=143 Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.321416 4767 scope.go:117] "RemoveContainer" containerID="fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.321619 4767 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kfvzq/must-gather-lvfx5" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.349568 4767 scope.go:117] "RemoveContainer" containerID="ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.405231 4767 scope.go:117] "RemoveContainer" containerID="fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d" Jan 29 16:04:03 crc kubenswrapper[4767]: E0129 16:04:03.405712 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d\": container with ID starting with fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d not found: ID does not exist" containerID="fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.405749 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d"} err="failed to get container status \"fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d\": rpc error: code = NotFound desc = could not find container \"fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d\": container with ID starting with fd7447cb80862e54f0bc95cacd4e765e698c22ebc94019836e316d65ce77791d not found: ID does not exist" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.405770 4767 scope.go:117] "RemoveContainer" containerID="ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd" Jan 29 16:04:03 crc kubenswrapper[4767]: E0129 16:04:03.406302 4767 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd\": container with ID starting with ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd not found: ID does not exist" containerID="ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd" Jan 29 16:04:03 crc kubenswrapper[4767]: I0129 16:04:03.406347 4767 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd"} err="failed to get container status \"ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd\": rpc error: code = NotFound desc = could not find container \"ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd\": container with ID starting with ddc5b51843f6e9d84cc7d2ac1a2b718c029a9e532a12c16a1e5ee9b1977141bd not found: ID does not exist" Jan 29 16:04:05 crc kubenswrapper[4767]: I0129 16:04:05.026188 4767 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5590f448-f43f-40c7-a902-672d79f0956a" path="/var/lib/kubelet/pods/5590f448-f43f-40c7-a902-672d79f0956a/volumes" Jan 29 16:04:12 crc kubenswrapper[4767]: I0129 16:04:12.806401 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:04:12 crc kubenswrapper[4767]: I0129 16:04:12.806969 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.706847 4767 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6lpxz"] Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707681 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707696 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707731 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="extract-utilities" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707739 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="extract-utilities" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707754 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="extract-content" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707761 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="extract-content" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707776 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="extract-content" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707782 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="extract-content" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707794 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="extract-utilities" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707800 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="extract-utilities" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707812 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="extract-content" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707818 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="extract-content" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707835 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707841 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707860 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5590f448-f43f-40c7-a902-672d79f0956a" containerName="gather" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707866 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5590f448-f43f-40c7-a902-672d79f0956a" containerName="gather" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707882 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5590f448-f43f-40c7-a902-672d79f0956a" containerName="copy" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707888 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5590f448-f43f-40c7-a902-672d79f0956a" containerName="copy" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707956 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="extract-utilities" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707963 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="extract-utilities" Jan 29 16:04:41 crc kubenswrapper[4767]: E0129 16:04:41.707973 4767 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.707979 4767 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.708113 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5590f448-f43f-40c7-a902-672d79f0956a" containerName="copy" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.708125 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b375cb3-2015-4d3f-b0d0-1e883fca2e5b" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.708141 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="d220a90a-2e09-428f-899f-9a0abc241587" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.708147 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5590f448-f43f-40c7-a902-672d79f0956a" containerName="gather" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.708157 4767 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7003ac-5dc0-4319-92e8-203585c220c9" containerName="registry-server" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.709092 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.736962 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6lpxz"] Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.851281 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c498c5-a313-48f9-9689-f1f97c5587ab-catalog-content\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.851346 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4tmx\" (UniqueName: \"kubernetes.io/projected/22c498c5-a313-48f9-9689-f1f97c5587ab-kube-api-access-k4tmx\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.851406 4767 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c498c5-a313-48f9-9689-f1f97c5587ab-utilities\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.953320 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c498c5-a313-48f9-9689-f1f97c5587ab-catalog-content\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.953383 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4tmx\" (UniqueName: \"kubernetes.io/projected/22c498c5-a313-48f9-9689-f1f97c5587ab-kube-api-access-k4tmx\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.953433 4767 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c498c5-a313-48f9-9689-f1f97c5587ab-utilities\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.953816 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22c498c5-a313-48f9-9689-f1f97c5587ab-catalog-content\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.953881 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22c498c5-a313-48f9-9689-f1f97c5587ab-utilities\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:41 crc kubenswrapper[4767]: I0129 16:04:41.971336 4767 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4tmx\" (UniqueName: \"kubernetes.io/projected/22c498c5-a313-48f9-9689-f1f97c5587ab-kube-api-access-k4tmx\") pod \"redhat-operators-6lpxz\" (UID: \"22c498c5-a313-48f9-9689-f1f97c5587ab\") " pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.031346 4767 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6lpxz" Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.459627 4767 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6lpxz"] Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.619061 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6lpxz" event={"ID":"22c498c5-a313-48f9-9689-f1f97c5587ab","Type":"ContainerStarted","Data":"66c4f8b08d80f46a788e7fccb372d2b74e000b1872b58006763f13d6b6ba8ee2"} Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.806120 4767 patch_prober.go:28] interesting pod/machine-config-daemon-kgsln container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.806202 4767 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.806257 4767 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.807084 4767 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2"} pod="openshift-machine-config-operator/machine-config-daemon-kgsln" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 29 16:04:42 crc kubenswrapper[4767]: I0129 16:04:42.807154 4767 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" containerName="machine-config-daemon" containerID="cri-o://2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" gracePeriod=600 Jan 29 16:04:42 crc kubenswrapper[4767]: E0129 16:04:42.967986 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:04:43 crc kubenswrapper[4767]: I0129 16:04:43.629747 4767 generic.go:334] "Generic (PLEG): container finished" podID="c144460d-d956-4e9c-8354-bfd72b970e30" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" exitCode=0 Jan 29 16:04:43 crc kubenswrapper[4767]: I0129 16:04:43.629815 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" event={"ID":"c144460d-d956-4e9c-8354-bfd72b970e30","Type":"ContainerDied","Data":"2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2"} Jan 29 16:04:43 crc kubenswrapper[4767]: I0129 16:04:43.630205 4767 scope.go:117] "RemoveContainer" containerID="55581c47200bca1e0d0702088fd9033cee9ab992eba215fd248062d21ef80f24" Jan 29 16:04:43 crc kubenswrapper[4767]: I0129 16:04:43.630752 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:04:43 crc kubenswrapper[4767]: E0129 16:04:43.631016 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:04:43 crc kubenswrapper[4767]: I0129 16:04:43.633557 4767 generic.go:334] "Generic (PLEG): container finished" podID="22c498c5-a313-48f9-9689-f1f97c5587ab" containerID="4bc1046a18142ce796f8712a9167dff6f582fec951ec28574aeba14ca3b0c076" exitCode=0 Jan 29 16:04:43 crc kubenswrapper[4767]: I0129 16:04:43.633603 4767 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6lpxz" event={"ID":"22c498c5-a313-48f9-9689-f1f97c5587ab","Type":"ContainerDied","Data":"4bc1046a18142ce796f8712a9167dff6f582fec951ec28574aeba14ca3b0c076"} Jan 29 16:04:43 crc kubenswrapper[4767]: I0129 16:04:43.635966 4767 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 29 16:04:43 crc kubenswrapper[4767]: E0129 16:04:43.764254 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 16:04:43 crc kubenswrapper[4767]: E0129 16:04:43.764403 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4tmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6lpxz_openshift-marketplace(22c498c5-a313-48f9-9689-f1f97c5587ab): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:04:43 crc kubenswrapper[4767]: E0129 16:04:43.765938 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:04:44 crc kubenswrapper[4767]: E0129 16:04:44.643340 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:04:55 crc kubenswrapper[4767]: E0129 16:04:55.162214 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 16:04:55 crc kubenswrapper[4767]: E0129 16:04:55.163140 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4tmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6lpxz_openshift-marketplace(22c498c5-a313-48f9-9689-f1f97c5587ab): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:04:55 crc kubenswrapper[4767]: E0129 16:04:55.164316 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:04:57 crc kubenswrapper[4767]: I0129 16:04:57.019083 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:04:57 crc kubenswrapper[4767]: E0129 16:04:57.019779 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:05:08 crc kubenswrapper[4767]: I0129 16:05:08.018318 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:05:08 crc kubenswrapper[4767]: E0129 16:05:08.019092 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:05:10 crc kubenswrapper[4767]: E0129 16:05:10.021448 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:05:22 crc kubenswrapper[4767]: E0129 16:05:22.160693 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 16:05:22 crc kubenswrapper[4767]: E0129 16:05:22.161328 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4tmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6lpxz_openshift-marketplace(22c498c5-a313-48f9-9689-f1f97c5587ab): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:05:22 crc kubenswrapper[4767]: E0129 16:05:22.162472 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:05:23 crc kubenswrapper[4767]: I0129 16:05:23.018494 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:05:23 crc kubenswrapper[4767]: E0129 16:05:23.018991 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:05:36 crc kubenswrapper[4767]: I0129 16:05:36.018137 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:05:36 crc kubenswrapper[4767]: E0129 16:05:36.018844 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:05:37 crc kubenswrapper[4767]: E0129 16:05:37.021880 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:05:49 crc kubenswrapper[4767]: I0129 16:05:49.018800 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:05:49 crc kubenswrapper[4767]: E0129 16:05:49.019635 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:05:52 crc kubenswrapper[4767]: E0129 16:05:52.023203 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:06:04 crc kubenswrapper[4767]: I0129 16:06:04.018641 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:06:04 crc kubenswrapper[4767]: E0129 16:06:04.019390 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:06:07 crc kubenswrapper[4767]: E0129 16:06:07.171244 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 16:06:07 crc kubenswrapper[4767]: E0129 16:06:07.171858 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4tmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6lpxz_openshift-marketplace(22c498c5-a313-48f9-9689-f1f97c5587ab): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:06:07 crc kubenswrapper[4767]: E0129 16:06:07.173464 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:06:17 crc kubenswrapper[4767]: I0129 16:06:17.018673 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:06:17 crc kubenswrapper[4767]: E0129 16:06:17.019393 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:06:20 crc kubenswrapper[4767]: E0129 16:06:20.022178 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:06:28 crc kubenswrapper[4767]: I0129 16:06:28.018155 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:06:28 crc kubenswrapper[4767]: E0129 16:06:28.019931 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:06:33 crc kubenswrapper[4767]: E0129 16:06:33.020769 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:06:39 crc kubenswrapper[4767]: I0129 16:06:39.018626 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:06:39 crc kubenswrapper[4767]: E0129 16:06:39.019352 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:06:44 crc kubenswrapper[4767]: E0129 16:06:44.021710 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:06:53 crc kubenswrapper[4767]: I0129 16:06:53.019088 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:06:53 crc kubenswrapper[4767]: E0129 16:06:53.019920 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:06:57 crc kubenswrapper[4767]: E0129 16:06:57.024024 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:07:06 crc kubenswrapper[4767]: I0129 16:07:06.019157 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:07:06 crc kubenswrapper[4767]: E0129 16:07:06.019967 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:07:12 crc kubenswrapper[4767]: E0129 16:07:12.020952 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:07:19 crc kubenswrapper[4767]: I0129 16:07:19.018033 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:07:19 crc kubenswrapper[4767]: E0129 16:07:19.018705 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:07:25 crc kubenswrapper[4767]: E0129 16:07:25.027601 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" Jan 29 16:07:32 crc kubenswrapper[4767]: I0129 16:07:32.019325 4767 scope.go:117] "RemoveContainer" containerID="2271a6153c58c5b8f857cd9ff44129da89b67bcc0c4ba9c0ff32bc4e6832eff2" Jan 29 16:07:32 crc kubenswrapper[4767]: E0129 16:07:32.020140 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kgsln_openshift-machine-config-operator(c144460d-d956-4e9c-8354-bfd72b970e30)\"" pod="openshift-machine-config-operator/machine-config-daemon-kgsln" podUID="c144460d-d956-4e9c-8354-bfd72b970e30" Jan 29 16:07:36 crc kubenswrapper[4767]: E0129 16:07:36.160702 4767 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 29 16:07:36 crc kubenswrapper[4767]: E0129 16:07:36.161238 4767 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k4tmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6lpxz_openshift-marketplace(22c498c5-a313-48f9-9689-f1f97c5587ab): ErrImagePull: initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)" logger="UnhandledError" Jan 29 16:07:36 crc kubenswrapper[4767]: E0129 16:07:36.162440 4767 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"initializing source docker://registry.redhat.io/redhat/redhat-operator-index:v4.18: Requesting bearer token: invalid status code from registry 403 (Forbidden)\"" pod="openshift-marketplace/redhat-operators-6lpxz" podUID="22c498c5-a313-48f9-9689-f1f97c5587ab" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515136702732024453 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015136702733017371 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015136672423016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015136672423015466 5ustar corecore